var/home/core/zuul-output/0000755000175000017500000000000015134211121014514 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015134220416015467 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000220203615134220360020247 0ustar corecore qikubelet.log_o[;r)Br'o b-n(!9t%Cs7}g/غIs,r.k9GfD P}FYI_翪|mvşo#oVݏKf+ovpZj!Kޒ/h3_.gSeq5v(×_~^ǿq]n>߮}+ԏbś E^"Y^-Vۋz7wH׋0g"ŒGǯguz|ny;#)a "b BLc?^^4[ftlR%KF^j 8DΆgS^Kz۞_W#|`zIlp_@oEy5 fs&2x*g+W4m ɭiE߳Kfn!#Šgv cXk?`;'`&R7߿YKS'owHF6":=3Ȑ 3xҝd){Ts}cZ%BdARO#-o"D"ޮrFg4" 0ʡPBU[fi;dYu' IAgfPF:c0Ys66q tH6#.`$vlLH}ޭA㑝V0>|J\Pg\W#NqɌDSd1d9nT#Abn q1J# !8,$RNI? j!bE"o j/o\E`r"hA ós yi\[.!=A(%Ud,QwC}F][UVYE NQGn0Ƞɻ>.ww}(o./WY<͉#5O H 'wo6C9yg|O~ €'} S[q?,!yq%a:y<\tunL h%$Ǥ].v y[W_` \r/Ɛ%aޗ' B.-^ mQYd'xP2ewEڊL|^ͣrZg7n͐AG%ʷr<>; 2W>h?y|(G>ClsXT(VIx$(J:&~CQpkۗgVKx*lJ3o|s`<՛=JPBUGߩnX#;4ٻO2{Fݫr~AreFj?wQC9yO|$UvވkZoIfzC|]|[>ӸUKҳt17ä$ ֈm maUNvS_$qrMY QOΨN!㞊;4U^Z/ QB?q3En.اeI"X#gZ+Xk?povR]8~깮$b@n3xh!|t{: CºC{ 8Ѿm[ ~z/9آs;DPsif39HoN λC?; H^-¸oZ( +"@@%'0MtW#:7erԮoQ#% H!PK)~U,jxQV^pΣ@Klb5)%L%7׷v] gv6دϾDD}c6  %T%St{kJ_O{*Z8Y CEO+'HqZY PTUJ2dic3w ?YQgpa` Z_0΁?kMPc_Ԝ*΄Bs`kmJ?t 53@հ1hr}=5t;nt 9:I_|AאM'NO;uD,z҄R K&Nh c{A`?2ZҘ[a-0V&2D[d#L6l\Jk}8gf) afs'oIf'mf\>UxR ks J)'u4iLaNIc2qdNA&aLQVD R0*06V۽棬mpھ*V I{a 0Ҟҝ>Ϗ ,ȓw`Ȅ/2Zjǽ}W4D)3N*[kPF =trSE *b9ē7$ M_8.Ç"q ChCMAgSdL0#W+CUu"k"圀̲F9,,&h'ZJz4U\d +( 7EqڏuC+]CEF 8'9@OVvnNbm: X„RDXfיa }fqG*YƩ{P0K=( $hC=h2@M+ `@P4Re]1he}k|]eO,v^ȹ [=zX[tꆯI7c<ۃ'B쿫dIc*Qqk&60XdGY!D ' @{!b4ִ s Exb 5dKߤKߒ'&YILұ4q6y{&G`%$8Tt ȥ#5vGVO2Қ;m#NS8}d0Q?zLV3\LuOx:,|$;rVauNjk-ؘPꐤ`FD'JɻXC&{>.}y7Z,).Y톯h7n%PAUË?/,z_jx܍>М>ӗom$rۇnu~Y݇̇TIwӜ'}׃nxuoỴRZ&Yzbm ]) %1(Y^9{q"4e?x+ [Vz;E|d1&ږ/0-Vb=SSO|k1A[|gbͧɇد;:X:@;afU=Sru CK >Y%LwM*t{zƝ$;ȾjHim @tBODɆj>0st\t@HTu( v e`H*1aK`3CmF1K>*Mk{_'֜dN${OT-n,'}6ȴ .#Sqη9]5zoX#ZVOy4%-Lq6dACYm*H@:FUф(vcD%F"i ' VVdmcOTKpwq.M?m12N[=tuw}opYG]2u<ΰ+a1tHayɒ aY(P*aaʨ@ΰ<pX X{k[%Egl1$9  ֲQ$'dJVE%mT{z`R$77.N|b>harNJ(Bň0ae3V#b,PY0TEu1L/]MTB4$`H6NI\nbǛ*AyA\(u|@ [h-,j7gDTÎ4oWJ$j!frH_HI\:U}UE$J @ٚeZE0(8ŋ ϓ{BpY]Q4`Iz_*2coT'ƟlQ.Ff!bpRw@\6"yr+i37Z_j*YLfnYJ~Z~okJX ?A?gU3U;,ד1t7lJ#wՆ;I|p"+I4ˬZcն a.1wXhxDI:;.^m9W_c.4z+ϟMn?!ԫ5H&=JkܓhkB\LQ"<LxeLo4l_m24^3.{oɼʪ~75/nQ?s d|pxu\uw?=QR -Mݞίk@Pc n1æ*m$=4Dbs+J \EƄզ}@۶(ߐ/ۼ𹫘qݎt7Ym݃|M$ 6.x5 TMXbXj-P\jА޴y$j`ROA"EkuS#q * CƂ lu" yo6"3껝I~flQ~NCBX`]ڦÞhkXO _-Qy2$?T3ͤEZ긊mۘ$XD.bͮW`AީClСw5/lbl[N*t*@56."D/< {Dۥ sLxZn$N(lYiV =?_e^0)?]{ @| 6+#gPX>Bk2_@L `CZ?z3~ }[ tŪ)۲-9ֆP}b&x Uhm._O 4m6^^osVЦ+*@5Fˢg'!>$]0 5_glg}릅h:@61Xv` 5DFnx ˭jCtu,R|ۯG8`&ו:ݓ3<:~iXN9`2ŦzhѤ^ MW`c?&d.'[\]}7A[?~R6*.9t,綨 3 6DFe^u; +֡X< paan}7ftJ^%0\?mg5k][ip4@]p6Uu|܀|Kx6خQU2KTǺ.ȕPQVzWuk{n#NWj8+\[ ?yiI~fs[:.۽ '5nWppH? 8>X+m7_Z`V j[ s3nϏT=1:T <= pDCm3-b _F(/f<8sl, 0۬Z"X.~b٦G3TE.֣eմi<~ik[m9뀥!cNIl8y$~\T B "2j*ҕ;ێIs ɛqQQKY`\ +\0(FęRQ hN œ@n|Vo|6 8~J[,o%l%!%tyNO}}=ʬ-'vlQ]m"ifӠ1˟ud9)˔~BѤ]һS8]uBi( Ql{]UcLxٻa,2r(#'CDd2݄kTxn@v7^58þ Ţ&VY+yn~F8I !6WB3C%X)ybLFB%X2U6vw8uUF+X|YukXxVO(+gIQp؎Z{TcR@MSRδ~+1æ|mq՗5$B᲋eY(|*磎\Dži`dZe j'V!Mu@ KV{XץF .Jg< ƜINs:b zĄu3=Az4 u5'og^s7`Rzu-anOIq;6z( rx߅ euPvIɦ7聀t>G;_H;2ʗ6 h6QװxmR JQUbTP2j˔Ni)C)HKE"$ӝ!@2<Bq 2oh80,kNA7,?ע|tC3.㤣TiHEIǢƅaeGF$ u2`d)/-st{E1kٌS*#¦۵_Vu3ЩpRIDr/TxF8g4sѓ{%w .ʕ+84ztT:eEK[[;0(1Q@ET0>@wY)aL5ׄӫ A^%f+[`sb˟(]m`F3 W((!5F-9]dDqL&RΖd}})7 k11 K ;%v'_3 dG8d t#MTU']h7^)O>?~?_ȿM4ə#a&Xi`O}6a-xm`8@;of,![0-7 4f kUy:M֖Esa./zʕy[/ݩqz2¼&'QxJE{cZ7C:?pM z*"#窾+ HsOt۩%͟A498SwWv|jNQ=-[ӓIgJ8@o2k'Hr~4Z(I8!H G8HNW%1Tќ^?'H(^jJ=䄸-m!AdEږG)շj#v;#y/hbv BO Iߒ {I7!UՆGIl HƗbd#HAF:iI }+2kK:Sov3b:1)'A6@\2X#Ih9N ̢t-mfeF;gUаQ/ .D%ES*;OLRX[vDb:7a}YF30H #iSpʳ]'_'ĕ -׉6tfЮ$zͪO_sYq+q艻*vzh5~Yy;,DiYTP;o./~^.6+zZFD& m@WXe{sa 2tc^XS?irG#^ŲDI'H_Ȯ;RJ&GT.Kwj;of¬zHmmS2ҒN'=zAΈ\b*K ڤUy""&D@iS=3&N+ǵtX^7ǩX"CA⥎å+4@{D/-:u5I꾧fY iʱ= %lHsd6+H~ Δ,&颒$tSL{yєYa$ H>t~q؈xRmkscXQG~gD20zQ*%iQI$!h/Vo^:y1(t˥C"*FFDEMAƚh $ /ɓzwG1Ƙl"oN:*xmS}V<"dH,^)?CpҒ7UΊ,*n.֙J߾?Ϲhӷƀc"@9Fў-Zm1_tH[A$lVE%BDI yȒv $FO[axr Y#%b Hw)j4&hCU_8xS] _N_Z6KhwefӞ@蹃DROo X"%q7<# '9l%w:9^1ee-EKQ'<1=iUNiAp(-I*#iq&CpB.$lٴާt!jU_L~Tb_,֪r>8P_䅱lw1ù=LAЦz38ckʖYz ~kQRL Q rGQ/ȆMC)vg1Xa!&'0Dp\~^=7jv "8O AfI; P|ޓܜ 8qܦzl5tw@,Mڴg$%82h7էoaz32h>`XT>%)pQ}Tgĸ6Coɲ=8f`KݜȆqDDbZ:B#O^?tNGw\Q.pPO @:Cg9dTcxRk&%])ў}VLN]Nbjgg`d]LGϸ.yҵUCL(us6*>B 2K^ sBciۨvtl:J;quӋkKϮ듃ԁ6Y.0O۾'8V%1M@)uIw].5km~Ҷ綝R(mtV3rșjmjJItHڒz>6nOj5~IJ|~!yKڮ2 h 3x}~ے4WYr9Ts] AA$ұ}21;qbUwRK #}u'tLi'^Y&,mCM)eu㠥Ѻ\a}1:V1zMzT}R,IA e<%!vĉq|?mtB|A ?dXuWLGml?*uTC̶V`FVY>ECmDnG+UaKtȃbeb筃kݴO~f^⊈ 8MK?:mM;ߵoz+O~e3݌ƺ(ܸf)*gCQE*pp^~x܃`U'A~E90t~8-2S󹞙nk56s&"mgVKA: X>7QQ-CDC'| #]Y1E-$nP4N0#C'dvܸȯ.vIH"ŐR ;@~y>Kv{) 9AG ćͩ$.!б~N8i"1KФ\L7/,U@.ڮO?mُa ې!rGHw@56DǑq LA!&mYJ*ixz2*{_;IYJXFfQ* 0kA".mݡ"3`Rd1_u6d逖`7xGMf}k/⨼0Κ_pLq7k!dT x삖A7 u/~&ӄMu.<|yi I?@)XJ7{ޱ?Q]{#\4ZfR-dVaz./f+yGNMGOK?2_~3\z=y}^G$*A! IcuR.o=MZ9zu b#s9@*иrI@*qQN||Ix;I}&ݢ6ɢ}{]x}_o>Mm8S]~(EX{nsa4$W 'ilΏf3kZc7l !¶6!=8?8[Y|-leǪzd;p-s~GM>e:9[v\:P 8'k01Q1jbX)/ΏL+ΆjBUx~Ga9Z"Q8_wjTLRˀtV5V:[g./0 +~䤗d7F >:֓[@ QPltsHtQ$J==O!;*>ohǖ6[|E7e۰ϕ>Uy`5pg/cc l 2Jj.Щ LkJu\!`0);Sak$Vfp~G8l WIfr\q4|UkC.`a@> zi 8& ׹U>8aK0%VR t!Lku`]c0h&)IVC)p| QUA:]XL2La& 1'v^ZZ4`9);q`F"dq1v>ժbLGd~MP5} x52LMF9 E*A,S V}\#X/2< 5FB΢u/`aJ#Tkš*D#cuCXȉ4 ՖK(KP|WdZ1?2u:|4F%_0E5IE'rX-|_N?;?V~ jh%޿| :S PxbFOor^3>Uf1w?a_V iэy(_άj]$9XN+/Sh]icc w0/ %IX: P/>J+Ȧcv/#]iX :)ǪCV..W%'X Uk}FflѓV Pwg Άj֘9bQ T5҃F<qbiRX%d˜u -ss x%| ݴ9V= %8~k ՟Sxa7 WS!EasUfB7™:%[Y ]LXg3۾4\.?}f [j·dMK!IZ5>H?wfS{]/oWxwWWWnoHKhkQ\|ZUN 1~u]ZTT걼/?Ykq{uu[^usqUKU\^|O-;:n'?ݸ>՟ 3[)O:n?埢&O8n!v߅l@Yd_nnzΙk0hAp̏a$X V)@6Wb uw:YɊ",1=1'ˠ*>qX/E b\yx)tb*H~%b@:Oec1^:V4FDza5H`:&Q5V ^hI8ʁuX EA>GЅ'JƖ'ZXO='kJՂ*]Dpa)% qZBj𒄓(~ |ؑ3,$ *>meيO>Y?> (<2QhEB- |q.ȃTs| w&5fGQ %8Gk`N;SlRh)5؈Vx1Bd="kԦ:,EVewN ًs9d#$*u1A \5~n#^*˟ R (0\lCulɱĘʦ|k*+ŘOi й qPq3T&mFqZbRT1e8V SncȱLȍ-03cu0:U[tp^}{~YhilS&Z!֨řoҚ*HKX 6„=zҌ5+P1;և6UE@Uo/18.fg7_]k]kDΡ1R0/!_Ɩ} 8& Ў3)X<,se7fS;W{@dlG)%OعF&4D&u.Im9cH,HCԢ[b p@-LG+@_$c%* jRr\:dc5u= A`U1cJ0Gk@5^NtY"FU#X) b0G'+_1&1 7ɜc:xF 2侯}>Ir~Z#+A<й l~R9zyƖݗ&!jC`@ qэ-V Rt2m%K6lX)"Clj齔GY:VmSk}qc?Nm]'`#&<6JJ4U88A1gXey-,)i7Ո: tC\?c%7\9:W_´5 >Ę_"+BLu>'Ɩ=xn[⠋" ѫQvPlp;3F$Zveju|Ɩ},X n*l[hq\A2) ' RGvh?>)-&188 \6y 8V p-lqrG]?~/3jsqE~ sjZ+9[rAJsT=~#02ݬ,f¸nj,@2"?WYؾNr<V`=V[B5!ZZ\ļXǪЎrce1P[;qsC.č-]W‹y??ڍ={6ݘ)|Ը>U{ѹ .MVfzd0O?7- }|>\T%9dp -*Nu6Y27[yBPj|Lc~L{rQۻ/u}q}hn+.kpWEqws]]|-ǫ\}-J.MLmc ;7ԗWrU}-Ǜ+sXn[>eywyY]]¨Kpx c./mo;_oRy*4݀wm&8֨OreXMlk)=VzpO֠24h 1hi D{q>~%̈v^nBi~MefZN >:/??ǴTߧcr_$dts`3 P}Kf@r3 Zj/>||V ?\§&+rt>g~j2H`R<(*2jF_X:,o^$M3M-"uo&ns&˿<*1>r,|2.q<(G3E5EZS"/q&xO}sWc@<8ϐ2,o<%e'FtJNؿ8u 2t #xn|L#t_+c/C=i>/Eyq嚆I}\ݝpC/x^\Y!2WDKt 9% 5HWfڀ1ͯ }c1x\= +Z]?rAż!3q|9oPdX֋|Z\Z rD$` <}sÓDֿ;<&"3ҽ:vu8B7g[1ӏig[p@I?C_<}}4ǧ4G?'}OCBztpzkC e5;Msq}' #yʋt{uൺޤB_ s3_;gө'dڟt A_e=jAa+ `cM6]b.!OoxDa==7;t-ds"L$?T讼F2Cc9a=9KdTD"F,M`W"I䌞vѩP uhw˝k֚HrOsXqѲ'Myrr_KqrG*A&AOockd]aEd{xQ,d)Ϛ*//X A*)lNWg$09GߌꧼuÓ_ި%zy>(IkThxsfvycZ{yVN.N=NQ0 8ox-0_.q]6; &T|Nx<[վhpD(lo(oO4C=Meޠ~犗R@$ϳ>/j !8 9D_r<h !Ǔ7M.XWۆ$6,Ox>VB2 ~û3u{0CswCBCOf$l8 Eq$9:iB`']^h@_y:)eE93\۲򦖯;zV4P< gFms0P!pmzwoyZ$2t>dod3g^/heoۀ7`NHLs?/j#t;fmv2i\,Zu] XOs'cys4ߟtvABWVѐܷw]=)M$Hy/t.dؔA?z y癄a9;KE*`Nx ߿䓮Zƴā {WU9:ꅜcFY/R4N7g`h`,?8Y4|pa߹>I֭7qŃd"yR/N2@hjsx7'$ c-5U#T.y{ceG16 j,b٦CRЭ3TVOJ{xr,aѶ߃'X/2d7eH+`=O%HDD9a^r90 ᜹;!̛"EkWUצ Aq-=!!Z$6#KB\&lz]#a_(Zs (o!u=}p 6 k]9C>;pTl-'q )0nyh@~>||!z]ɀBi6!\Ft]}ö$|^!{ӻW?hGfH|/$ɯКċaѵTt%qGNf^)=:v%E_E<"ph|dzED⫑@o࿏^%WzE#lVYqc4]<b`XwfE3 KԝJ<>łe|!(U FH7.n?*+|eL MyQ2iӍ GDR.* ~O^ ئ߿d)~15 h.>'ST I=j`#gQYU ?'"` c))Ͼ,鰞鐋e\mOSD@p\ʯ} ]=n3gl%DjrYX## Z769U pKLx*">aFb Yў|P)=~haTiǹhs"ԄKC"XZV+Lcw#!ص m*\$gռ MHkU̝ 49l(mOsY BDo+y!gs|ڞ`2ȭ*d 0 [65Qi) +kB*) )< ꑐoH߽R R0'%$v`TUMJpSJ.yV{=ʔ;D}};Lր0ˍoylo%VZY^ʵT*li{:apıL2Ӎ^Yu4 ̿5@ ԼQ5̵A&4}5)Sav`mRyVy䶏q}UN3s.8BJfLk*x{;jQTX*GxrQN;҉)W`| !hCu ky]ϳHypMrU"lW@A 9ႍkEx|ASm|@T.H\*Id693T!Iʳ8U5eBbv"Z3GETj\]MgKn ?DLMVA$(j=ʚA#{nm_(yyٙ{13|ՖbS#Xb[ 񾄠ᶩS/2SUhcARSN HہhK{9; ls⡯ouOgrD~ jmMFTT9MM!x-7CKjuTEyjU'\?*F XPJq}|OoZZ)? s9B[S]xX.ZpZcJ~qfyl'u4Ћ|3.8̳BƸt7i!Jxu4A?b[&5tgzʛZ_n~ڒ U(P m*. OP&Kꐖg* ƖaHSX#*\H$!:]sWF7U<{}~$]A%詎U-S[OZI\kGb]DO%mX@tVD} t51G9ѷ|c'mǯ@^vY8HVݛ^_Wg0iiwmD n@ڨ3|7AU0bX "ws8tBp$;TG0ުf|dq[huFA$<08>jσ/\YwrH} [{ caQ[ܘL !TXܣKUG!X3ǜu!{HI:yTU7YU#!STu)^Üڪs;3ͱ Y,d*Sc}־aSew so:8Y}ob1vyӪ&n##ӳlC$a_ipE" R !UPX:/$ DsTu=@zʀc9˒4L"Zk-:iꚩI?Vjq#3{Kæ:Da'?ka7j{WmWΘe(19I;*2%@qzl cI} sbcޣC7d0Deʟ1JMb=uzƾpGޱO:y(uf):|q/2/O`@~t;#4dd@p@o<S֠+R71gѶ3ț5Z@yM]qE˶zŒ @d@AL !W`[MS`q^ [O:n>@|۠ }>%;4ɦD^WޝaGTt>vG$0y O(Z4m5hztΒY#kBsH޿{Y9єq^?H|fgX\#ĚYw΁hO cyРx?]tIAwϽ&"L:ﮀu?9$Xg1/WbBJA;q+M! @30B#\=nZ޲Ұ8ց8(Îu擋w-mZ`8s;w>ގNEM4B &=~;MkG0:{ҟ< cxH:U[:ķ3Ue?TiyI(}|8njdU3X @]u4O_>Px료!pDIkK'm׉nmx]#_gA^{}t`;:ޟG9F;y/"@ѰBpٯ 0v (wÅPx<κn ;Nw hU4h 9 a}k0NYgQt101ݛZ׷Ay]RXO%+N7E7YBC`1^(  5 DCm b# шr|prDtikn=Չ]^F ~s6p|V\QG{/DL E@8n׼9_lH94B,=y1E<`q t#C l5]sh9!^ hvxQ>]TB #Ն@nF;{ }5!q/\~$ķ]w饻B0F62$.&_^_\WF,4ewΒ :WЎ~(ʸp{c;) % k~T0#1s!=@u$1@$uOyĩou˷9]7F % wu)?3s8gV7L됡3<;2>+ARۑwt} LÌ[?242>K8g9Ӱ`] f̷~X=zbp7׉3κ&Ǚjnkvvm syVq[#qѮ]oߑnk~#y{Fq[Cz68rHxWJ\NTv[w],I#gȮLxp[CUd6V@NdF AgB$A7B!O ]pCC93$(||c޳(;Nl8:*bje9 }nv$;jar0iǫ/ο:|pE˲H 'b3)Rd)^&^͠v܀0\.K`gDŸgiG8hepX[o$Ċٍ஑= <WQyѶċu!mG_szKQVQYL18TnƂf,IOӉ׀ <.0:K!FUr.7W*W#%#Xg`M3 H נā^$:x][ς721gZde!4 z 4J ϸ&g,b`-Pڋwƅ3AL=#9R\z &#8a` !GpxX7`m&RÒ*9pLq9{xyS|p!MB^e`Ӳ<0`Cop,*cijv%V87)$򼀅vD(>`uu ǀ|I>Wٴ ff'L6zp@ρ.]W$ xY<%vЧ]O-&j 8 Σ*M 2.. Jj GSGT }̌X$c\SRJ1F u*G f(j6XjbpP)U@eJ?$Iyj.Cn`7ё{bR,\и('>)tkh^Ef̼tC;LEA2~8Hu9.CMFMꛮeRiUi'7r r?dHЄ ib>m۟tH_rIwHl \/_M» ֕%(",O'd:N"B%Z_IKrM(10M=闙gVjqZ`ѵ<\O(?PVZܺWCpm*pTA#*sp ZeeVZe'0˫Y/Z'B"X}ïip, v>9BG:Km۸R8_.3 ,֢tRI2mυ݇a&<7? >DzrNCàt"oviq$ ՟)*(St =AwUo1X4|+gDa@mC *4&b!%0tH={ԅȌFCόg%Ѕa$P֏⑝x9r(,>s̡9fQ ={ 5/R ,VfeiyDP R`5 1!F]U<kpohY4203kC#nږT2p"_5\$&  ΕvgDxܵV'_aeLșoMg17!w躡'T OB`)YRe9;9eWcnկ;v}L>oݹ.[mA,۞V4R#)e[j/joA=I2:ۧN2]_8/w9_=<-UM3ޟyLˬJɛ,GwġA;}4([ZkZ W"ѕ gpWz6hz1ecahEB1`Z U ȁ<}p'NP ~Y.:v[/n1Pv~QR`9 QDi˟dhH|\:]Glf GhE|O?xF0CeNhT4wsu8 o _8=J:jR(Ӹ-Z\#49EԀdXڲ"[nKI# \jmsXE@X-g<9_h>> RaLd+ Z$\u3^棢SW0gLDܬ~s@ ʈ`"Ջtrs;_57ɺ ,?S6R gV =ǵ҃NYHNzM2X{rva GEH bj{Y-Y sjn\vnpMvsFA x2hQ6leaI!S5u::Kv4]JD:(SZ gs]9T|@! ӁqM,>ߐ5r5~8Hq@,7+Y>W8*Pq[G(!k&}ƒb!x6p34L$m RB84ݩ*8Y`JhK0`Nd#tD+/7Zߢ:G\~ |mY0*|AZ2ej A 00 ,M〺[0ᇻ85.b>Oy'3 e }]WQQ9}8_f$J qHo ].,qH@d,XLnUL?϶v9, Wt+]V01,?i tFrEDWqrJHx2o~c@\XOl%KWEϢD*<-iujz6SS) W.[:&_GANHnRY*XӵB8BkY6Ϸ"pc97EŔOhJ2m`(cߚ Aw L&";DWE53G%O6*KK6LNNR7Q`0*CTX aۓhI*bD!P3;ͩ.*H@u)y<[%Ha(ǵp([S ҳioMR CĭF扈2H}V;@k]#UtJfP}  9 `L^+ EGF#na$[jSzG6 7ѐff QǁpuNHˋ;B Nȣ<#u اK0^ *1̳p^ok8c$P~L8%˓^*_CZT|zg˛l{S8p7b3|{']2m/޹%LɴwXa<+4(l$1=7*9BАۢD㞠k7X<- zb G1X"(`w&i@ZfY9[RE9Q]wPf$k:闘^ "[Q @Y(נ|gf_))(9;3t02T=оD_bMfHoWl|ާU3YEB/"Z-z|]ѦZ,F4ڛH51h>k!jZceDa_ GK޵-q$_QVu8b>`vϳ+ ($%}f5Mh`fz%ZJtdVo_A# t/JBӠls8Z;Ad|wwp y7%'ob$i^bOlPopZFX`+KUYfl me3Èae')G|3g!+- X.Km7UX[c1I3^`ö]Yoj^OOI;qykc zĽU#'xI'<)mt_'ME8\ZOA;t16Pƈaimf찦\D4C/7fZ[;8̴aUeyv[s1.+48ΆpIqDoq|c\E=sWqs@cJKٚc;5qusM1mc$YWf "C*Ȫktbd[*Dk:oCv}ehNLR1]Lї"=? rbY=lġNjR,0ڀ:$v 3`̢R#a'YooؑY c.9YvA3]g!@89 'Dj!Ftaz}9lؼM(FR3KQet Icޒp3],6&mQ-8?#"=I<2i+8k2At,PrFIbİUYZ*{x\˿& ث2@!ZZ=WBYشwY"F#Y|< ȉV 3R ΪT gVdr1칦nriNDz:hUi dv1EeZ|Q"HHߚ$8NJ=4"_"W])"F'#e0Az V FKnwpCT=E-22H"<58pwQ_/H4wJ bʆ9QA*Kٻa׏3IwOIc{P3sAp,uM '|2jQ- sѢƁ!AКE!xF;<jpK(p14xFɴ[8@ qօ1Fhc<:BN|#~oKZpEhp%BF䢵ʤHA*zC߄5d聐ĎkIc!]eJԬ4PTթnlOJZ# B$BZGTi'S1Aw)a[6ǀ}1>XSrJi!x؊ W8p8b$ԧ"!BN Mk{SC4Ƽ) w٩AݪЄzpCѵ͖3C[5Ip aL0$ofXiI8J6; ")w1\a:G AILimuiIS1_]F (%~%v IxK@kp3lKX!R^TR|oϿ|Hp=E3D:cT?ńX_V~ҧYuSvnZ5.4Ϫ U[8R87Zp! 9uta-~~ QE]4gQ[#QE%bgmD,݅ ˢNN"k8oy}HbXGG9;n,D{ېaYyޏ' F+G9:pUJ슩'uMt'% jcS:ˤK1I GK}Z"qĵ3NM}JRƽ!ٺfzح򞬥nCEL=7(YNH4W7Dwvpߚ9BAJp35BhQ:sL AU6Fܬ6_&HIco[u>j}Ԫb2XZ $f2}ޭl"q\)"|rkd7֪' bAIߞ$5kuk,lEBHs{Q]$j0t+F҅* N繻, ̟Bh۰, Y5ˆ%Oѓc/ 4d'ƦZ6ppZt]Xϛ+t#钀!<ߒ({'R%2c8,fDa=%h1\6V>TB>gq O\w0I+=msO?c /5d*#}s3)[e +=XTF } ?r@],T]&(ECV܏ءλW6`9U;N" |Ъ;mH^X8dB­)iY10B {.޼?N5C[tp<Nw# Q}2.r'%\:xKN,jv9jO0oW=XrY.4Δm1t,$!CH0yޑϕkFEY }JӠ_s< F҅Z/#њEm 8Zp1Ϣ]n$8:U*tWJV,ߓâvϦ0k/Z#zOV4&SX`!Q=ӨQ9Go4J'"74X%8~[T [:sI1vQi;CFҸnԁ ;qÝ(`岶RUo(Uc+vOi =\#;GĂYJ~ rC1bVQ\ܔ;CԃIN t. |Q\v "yU$rVR/Tęj=7#@Ise- rYKO<6<)n9Cs; c5=DzXICxh~+S3G |P ŲC=KY ?M׶qCq؝弶v= ik;g)~LE_o$nG(Rh2T2Ԕsc] Wr#r|688zm_)WIOuC?O|xx%S5޵qce |?[vkl5CS>k%9i65HHvȶf<;琇$k?Y*95+ʪ|V%3өwcR1e'EA;?f}p, IN<@ݝKf2ƕn?pDȭ rxRL8cLb{vzDZ\EQYcB2@ {pyhV׹D̹̏HMUkk!}:(1h>簙om?}?qދe,hib/c<!:ꜞcM$R*nJ"D"Mb]N˺ X}1̨sMDa^'@BW.Aw#GkfJ^3ױ?(֜^*Ly`:Ec-KQ { iva=>^vK{ӛ^|fp[{oPe x]{YTC'`?PZ~;<,@}?AY0.J+&/{f2f?읋t4- @bSv㑇0AU;|Xhc9~]#`Cc#fI6L» ~dw )ӷepabp\i fj/nw%l7 @} Ds4D$Bv<.t)M^uՙᆙ3v-I7`h_qDAԾE0:YI_^ڛNb![p2|kBYN_f 83co\Ԕh$1`e$a? "6 3;=@#ɀg\Q];qldZfZbVg07J:8 ?[8?)ޝ,7 W.{qjMvի}[p06T]Qy?MUcU(~= \p}Sq;FmvL>6#iGfߟt^9l893A5E^=C̖'fcZw?n\M,uwǐ+?O@Iu]HxΏ Im9k֚3bܖ3VZ}+& %G0#@HLő d҃/71q@9IҀ-@ZE2" *H=aeJ+2-bMd`6$mPYIi{x2MZ(EJK9-TAVimN0R#cM䧴 \PmroD-m5G4FmeUl!T-b AuB܇ӾQm\I8W-W6 Qz-V{1+N6yy LB-3Ti" kZ׺\5g|! Jh7?t/zlc*KNO_K@bd/Gj'<8\I|ʔ : ~X} Ý9jU,c nF}AZ{dOqKjMCoyw6R\/N*u1ʅߟΒ&KN:NV%?PRCN F"9e@),Rt˘Ru|ûj=erԣU43Rd2(}VvҸ"w?8L(̓R\[)LZp~ܸ7e$ְ&C#&q=q Jw]ˉTXg!gUVj}9%PC}Y9\_Ku7 2[R2PgXKn68A cqYbYɟom*ߣxg<\ sFJGvNyUNaz-)YVs)4Fˊ6lEVƽ C|U#|` ifS;\#B$Q6P̽@q8\P}ڹK |SL74F=Cܠ lb烸e A#;\z#-/noGcmui 0tA-"5lk( Q,D;m,_P%pSZ`)[)ΣD@9)M} &vŷ ɭ3v8tHl&swl$W& Hk-pOazV mgJ]{ 4CHo#UmSP?Q;5^R7OFwggE6^^K(I\`kωcwl%!wO<<=h 6Mس;|&A{yCh7 ,+Īh(j- m߽%JvcE.Vޢ9Ū\;WkT?d9h.aPcx fWfR>5.^Rۻ qz_ѓ|+#l8W:B 𽳆 1\%ź.bB!2|Ի|[X]M:|"Tdק0e_DWWPBy' Q"xĘwGFcʩ̟{ski]i4?_dߜo1bӓ v Cs⼢!σF w<a<'sBsӉ8 qJ>懁q8]PU|rH>v.pHxIM91S_7S(`mu5GH}34S<D[EnMRq"Ym Rc,1KsK +#ނUae Xjf*7PFR"ĆdzdE[2;&'9&-E akWD 179b! 7 .Aij QXVE0o FDŽhv`)EX6ղ P"an(7P-6c9 )&X@3Mg{J1 7qrT_sr,/(&a sym)rgN)3{0 "gA'^_o<1,lkF0mv7z^\ݟ4fM11Qk1Rޮ bPs[ Mn9"8H+F%9wT,Hr`Cbb1tޖ44FS#YnɆ͠R b[6  V\v.\!&mPD6\KT ,>sd\@1qdkr3Ə{f2Y#5X9N{ً஄m}a$DK+-f+-B ~[ OcM< YK`": YO\ FkIKB "%QՏ{h`Pi[ޏ盾PqfBe~\TuZ ?u?FAR!#۸MUwe!ï_7~>g?YyWV-(wӯ]m/Ujw33}f_^w鱗_}q5.W-M<Gݭ;4`T3L#}?ꖩx\cFM 9ru^Z<4+B/k?. -}ғhLK߽aj~WN >^0ku% yafy+{%N.Dc +@yt|.JSgVޜaAcIcԌ+G*Wj u!F!#yɔŠZ3KnS7{R$גHhVxX7mEF"$qL9ɹ`@G(gӘ 0VTI08LP!) j7R gDR٢) ,DkB(m 8xF ?Pv KU[ u@.uM0hbfɄ<$0ĖE2U"+JE)*V&cD&n|rT:Ъ#oS[RJZl,>g=_zv Jf 7=GLb,I.nq !CB*o=\,\SebTZ שќI+$``S{Io}h HMRkX#9R3rbqH&v$`b&Nr`;`ˈ6c_ldncZfmT22i!81L3)i+ Z>BV3Bը;B 9%)@"0 GE.&1ŌO.wb) oN+1=\b*qwi*Jl$b\ !c91`biwuyYLMTJՒ\8KҁҾi>rRȝ6A>, 2pXV=l4+7#X0N؊l,S)H(VQ-Z$*nO)q0шRD¸JRBQ&$38QdQS#\llMQWI*Uiܺ}H igVXŚ@I$%Nb#ݪ`^;ٌ`qf*r\ވZ}LLetw#_ଅp `7}Eh;V0#7KwCHU$V䎐!c( RJ!4!Y)JD"Pu(1BH":_ESͺJ#-7"SZ灜fa%:oRk25SG?\n,򽰳a2ѱeK+q`jH-+%QPu`]nؕ+V{ % %m(lKgXG?L%ۼOdȱ&d*-ݛY{+nk/)F;X+kgٌIurT},֣nOmdA1oqL;cp~pCңf:z[4= Oeq?EP U34 $", N &Tr))̪((F&M ͳ`z08ax3,ZwW ׈kjy48pZ\_5OJa緁"l-[r#VMߒbxi4q_[LS$9KgY<d-ߺ#*{FJ < 2 #% `7𖛄-em=g봃%V6Xrs-3 ʮ#ʻ -Ct5X[zO{J5+Sh{ÿ= w AYaPơ)rMe6[7@o^Mm/)h\8^zOc0{Nbd"eja) ˍ=s qֆJ$AE649gNJqњP^8'ng ZR y91e$A}DК-;|ܥChb0Fdސc7">FPPj @{$_XA j#2ԺC*+;r\;r1?FDLoR72(ԥQg1y`RIqn)j]U DWHP DdUB*aTȤH? w9r%$-iPSҠtg "{-uIS˿vۥj{f^_&LfFcpyQGuadRZ$_K^⬯)Z|&Hg9~ș= ]7^zTJk6YJA%lj67߃y5LzSMVeU,/ \v`~isXab-8 RD j mB9p@K $ϨPMZsgl6*hn [פdRw{0/ `d4J~usxG"] =tf mN)ԓ)Hg  .'Kt_`쎖0~\|_%B_^QOڼ=}W#e^h5b y ϕ\o6 T*)a0䎔;\))d\A #u%W-dIe=p1Ľ¥)Ɖ8p1|h1v: Wp1cr+(bJFks+dz3^}>` !AEkOp勽jIlI"ţ5AGAy:xg~Iq3I.l2Ӊɦ؜:W|6P _ƾE{`JfW gJX:]H wQ%v%mgOZҩ۰Rv6oӑ<LDr/9"^x6EZl RڠW^g쑫ϟ٠%hCܰ\E/(Y<{_3%?p,^Y"SOӴt΢dc;a~),7nj :d>Ok1$O-)@O'Ò @0pXi~5N2O05$kt4^}zщm0wx/z[j?^ ,(.ۯ^҇ &KM인І!Z+zXDz w:R<֔|㏒J<lK`lj7*/q7h\MnİWW2-ο J}uaY>) ^<4*׭ ncSd2r7~n^5RTa[h#9N8L%!ȄZr&J[MX[ oc 5kL2}m>kA4pIjGcNQH)CiȤ͵RdIU>hZ1c@J SvV-)b,A:ya*$NU-dU!.v%*={CDoP(0I>mDŽFʄ\89\+Q.A)Ce5>\OJ`$JpD "ԧJ л,ZE˝<˦{7LTy+BK`kv V ;=p:نz,A$k-ך*S2E\eLUGeaB/e;:L`F$c?=^>4>B/[ʺ{U AhUtJ.E1((KS5EP\Sn; :hU1#X|ӫ:vYsOs7{!楐<%T$ƍ)ζZ=z}aQ>Wf#?T°2:R*d 0* c8SJSlC^57n{S61#M]S _gGe~|^NhRc6 z[Vx-C㨘ˣIq[d̳ ~7sW E.lKUrB 2{i6JUyxw]ez*W;t~9 ': ZHrhօdtYgYt=[8{=ȒQf ?ED,*qXlF"V2%F eIX*a44Dxb^X #*5Lٗ.ԃ:!4]Jl: Pd6wU2PdyYe_u _"0.*fM?vY:xl6h<86<ɾLb5ڿ6E#VY) CP]+JeMVyM+7˝^pugEq{m]/޷hdeUbYzﯓq49NfyNjQ"u(ɠqbPEA}f=uTîU:J`}ƄpHt/~:#Xp ^(K' U%@Kn͢y]vjY -eo1keĝ*ט>?x%huO-`[PP +.(69,g} שm=V\a ba /_/ TSFνSG<qxS<Ü+0$H_gtA[֋`8&Og 1:~J{,dS(P!Bf#ƔDHjnXBs ]bxw>8T$0iSdq. SDcmtUDD)cάI(I#NK3fFz=`Y#ݪ > y}kvBU{Qj+N`Ro_Ϧі']R*%;ŸW*y ]#("U39ŸW ^?DVb=⫰߳ԕ߫)e9YG,ļQ9[&޺徕6D*\;Wzhmm6Bnwm?,uE<-M(OLĥ f .li 3t:M@(j$MݕmOyNZz0ݺw+t߇Kߟ Dpq=SJp"a`תum탓.,9&(:ǮN)KcyrݯRgf{vJflzb :)"IcڔRiZ>U87z0=c,,^jCWgOC!_u! 4,਱ޑfGnp!7 IAp 2jp;-g%-o1_x[nzlRi~K ;W%˿]3*O/pAs IU g/+ Am+t%a!($WQ* -18j,ǁwe囶Gkl,ԍ j%V*KG{ fI:⊹EwdE*JUNi us"%9)Ͽ?Jydpb5+IkxW^ .{eklpp=w>-H"mկ !pMA5.@I[ w|'IlAQc8+kezՓz~dݒ+WɨmBskᚔ2%+)e(&}x7K*vү8*}mmL~l\*C O4_Ck m8 ]m:z[a$t6ڎ)A"9`.(qxOVӺux^c!ˇaF e$  $b2tD@ #pc(,ouaQ+a6E(&]T'~Js֣I( ܪduLwd~RبtQHV0xa!r<( VQrTe!M:xS3[Rx0[8M/CRUC18"Bˈexg,ixKiNzvB-2b)q. h~i[C>☵z$CRHB*ţ!0 FQk-%0&J%iQrxW7*wP#N4JA@F饑ƋЪp5pl+_?خĖZNy :I\2D W3ć yM_I1稵2caLSYT:KXc٢ ऍ;QF64,਱ޑہsJw @7]4e.^U8]`1h Ng, w{0!걊TP~`{TaZZkK')b)&fɑ& [c8q;T֙6J Ji֙\Ag9c}_bp-WEI1kbyyZ!XAixz92|0m/TuPK̈y[< xR;FNME "}p$&Kg}tU8yk܋#V'6戓v+zҴ5DU &@ļG4Шr \ sqƀ:WKh4 J:m4/y#L@ \Lߞ 4OW0Ͻc QQ"% `*q؛eњ.'1]:xS<@=cuX}AH4+n3?zE9 | pg^$].1U҅ y" 4k~͢턕/l6I)At"vQ:0yS A|\W!(L6وbyΧu#~Qυ *ԓ4kSwSSFK|C2QMuڻge_{^IBz-KT־#0j1MѾjc rx@%0G#E7uFAޖ!0MY_ xNw O蛦 ZXV`bqND8)f)$sZ`%z +SM3>cs384wum-'$Q~EpVsEn*QQO;\ B.[drh8E= F=R;/xt|[uܘ],z1rj'aR0װ1N(ALr{qUHa Q]2՚Pe6'bu#,}QJ31,0 iL8DK9$ ͣ"Rx 4FL$7>k$/'{yUq%l%HAB=f,qɍ@Q7Q]m~1>o z5©Õ% !5R w8iQ]ޕN7͕nh٥Q=K=eǷjô#i1K_T*݌OGG5O}1z1ʺ{J04D,F0 P^1y˃$#ė$ }@7FӺ]>E|c5{1+Z W}fy*Ye"$[liT^9G+Gwe.rKk~7TlHkzuIu^ ˌD-;]̧{5T/h*k Z.ϰ@ć-0mLhzDZ ޕBWy(M,^21lA J(KVFe"pԭ@xW+177cLGm8ic-s #r9c2Z_ҥ֕Ylюpl>/Z 4Fdv1mU`H-Nn `oS~KȩUNxቦV]CR& g>/ Z*AoB1X^UTr85k ŸBn$SI|ޭ.ƪT t ~Zz[=_u"~ שs闋ՠfq@6`n֎Dņ:[my x#> J|(!Vq Aٲ1|lņՈ/-M?n֪ԭnpw9 , *%>G5kF]vKC$ $q 'A3;`쒱XYwxȯޮw'U)O6Pn O*C Kٿ_mE-IsI5~IWz[UĬYZp א7?ybuj3d%sAkXfs?*ӫU zװ bLF) Uأ>dvf[R ?E(=b)fb0VBp;e@Uvݥ|O=N|4x K`-G1cLR8s D=HX``C baq Q=Tf(ḰARY=ĢJk%L  x+;Big" zSZ؁K)5Dٙ`K^5\r'okي?8>,446, bdyd p|8A^}tC# xmT̸nm`G׺Vu>U{%9p_T|F-˾ŋZTzTxNѽ5N@=A/q]ؚ8Je}AP*MMtzxGk>\G?/[UҀy "bFmG{9`9MQH60[pٹ{Łv[Jҝ+ʎ,(.E,!@إ*,U,=.qh @eyDu $)Έ'>>ek煹ps#.35s4&\CgzR5Q;1- @$׮ց[LmaTo 2 8>3Q˟jѠ3bg Zj-Wzkƫe^ TJWWcpXre-9p˛` Xj_Tc' R5^?įu·&A+ow=[OfuFOJV~:Kp3G(*W T5Oɀx[4^rC:^riAKč#6=`I5D7F9nxtocĀ.octfD;hL7, Mrk9e! ypx ;}‰u&1O}(lm:G{wte=7OF5*tsO:&Xjv>awʾ FՔXkD0=IM,y( Sp6k1ϵO'(eu&%_7.P:hL^\w+qlxUPj:c͑y}FnJ ?UbyBj"UjCnHq`+X*L|A18&2 |(I1TeةQDMvfc yj];+l)QISu&RQkrdɑD<ΑDeCbWp~M[w$Rj!KG͸3OP$Jv% |,xyΖOxjA-G2g x>#lf3݁F  |(YDRȀ™Lӌ€3XLCEq[Ȝth:G{OA!S1>PC*h*[U;hL2^vX2ؒXڐzdz;; tFes$m7M@<^qiJʡA(gt ӽL`͠ˉ5q\/VwN%12ћݿ^qEFT#ٳ~A%~tИrG۠z 4Qo#nQFPFY_6h/{rFNj2OM)g#_nŋ衊1S>dz0FI/ L^ѻ)AtKhb?uAcxWDm'nNq/RGD/y5'f>_>geR1GTڇkiԧji<T%ރ5j)8q ى BGB |(Y"x ѶN9,x<9]l*30jJ,(N-aJ%SV"I9Fֈ۶5 |(tFv= -xs(G`S.kIPO9)jp-aJe:hL8i&t:=1P}r1ل:JI\j7 VOSp#miTy{>t'3;z\[P}n#ZՓ."M BeC5}ruИ[/ZOJqw>O&QGlzaíx=!2XȜ"p Isb2Ey3-<1 ka! 56JٝĶFB0o$e{=^͠5 ɤ1&`5bQY4Jt'>Y \;ZM(! UdBK\$<ɦd#'x &"jZ/d57 OL~7[٠[{LLI}"pCV3<#Xu@ڶnPC%wIصY0__]ڱoӜ<*Dʿ_^^hTnq0Kij\R`BVm8JրnVTdzTI&Ki$, ZO 1 T%g5)`6JHh]c{}UD Cgig)-xdŒ_<֯3O,Q4=l ޔIkT Lۧ˗vPxSD{?Atdxgj^~5/w3z{k(`z3s:')w,+RMy6vPCLuIax3”JXcYN-פuVU`2>_֚`m]I <}<l36DIw")"Gul`[[ ;hGRG*\Bqs|uKp˂ߓr=`B q뾚{|SRnfn">Hv2_͎?\8I45Թ >wQr}JJ%{~Qu=(/SRim`QMrƹ$4No^0[_Nvv;w6ARIeryǟRzxK]XBίt+_=XG A8ci& @cӂ%t 2J`.,y(Ct. |\܄-#%'?p5|aat0oNz{$Խۥ93-*duۨL9o3&/l^(NA#2s8sڕ=SKW>#B4WҪ츒Fl؁VLU? vVp3Pszk5(@air@ Kݣsj_䰓g-Lp&u`wkSyJ]19hgc:bNPe="ibM0֥ 2iIaT_hg}̅ X˧a"] 1^(EH.of㗓ʽQr) 903 ]lw98^N[e jO]Pr}OjDH. ML2¸N[%>= j8/jZt}d\`/\S:@`%q3ﳡ#&5|{_Ԟ/yϥ9nbҔ$y8f LE#'nЫ{//쓊̭ 5{0}R)+ F=p V8{e$aW__]Kʿe;D3=9cGrS~(C~+lH.U/j6lqF~yvyuQ|)?W}R\+`(CRZ`wЈ̜t8s6>^.n72.vF5bNl/dԩ}QOyyTGSKLXFd(v{(.ws!χVH.CUpP L;hA:he@^ 0619O#VE(rIEuzh!`mTemO\2p%,mRR-G3:D\\}fv_*3VG5F0G2U6,26U:T aGK.肈ɉ8!^3eԞּSK#B4pM&QyIbyXKW﨔g`@p3oϕ1EmJN̗6*zLC\"9]"s" l*mْߑ\i Ayc{e&Fs>=E7[ib5Dcoq")ez㻒n_+Nb UtL'xM;biU%st)g}k K^WBz$+Wk r>C(͂`$:r$g{Nb|ൃꮃK#_1ݽqev.4Z.^0ҊKRPeH\v~%%Sn@$i+iCx l>Q+VZPηWʣ %i6v(JpX&)F2#XnPvFƤ[ ^#A056H֎:g<aЄf6U˛Ǖ`^hL߽{,(LS_v?;LT/_ӞI[ U}[/K7o˿+MhP9~ 4ًw] 5®/wq5d<ηt^mmeoFYF ͒^~K|ٛK`54yV<.v:|f.[-g{CcY5LY-x(N͟fnlk̷H]| ji~c|`Jw|bIJr딦$Ej ʭ.ʥ.y2MXeOp/nI60[Gn:)6PI*gbV‚cUxOJ;fn/C]w|7}ٔ7_ps40s쯏"~V~;j[Skhiڃy֦0o xU&t\ټ惮| *_mO\'~\lMT_À qUp+Y7֏|:iV.a[m]S0~eB-Au+ Fa__yU2t`s3o7~8]fXۀ{G,(> ǪjU~1a ?'-yKztX2 TfSX% 0y?@߬} +:AU هz#.ur1Pgw ^}뙭<1y`Ɨ.|N['u=p{pD [|Drޡ|B6[b5Ec]a\Å L@ GԦxj<$VjY:Gb*D7Я zj 0Va\C2m,.Q;hFkcʚر_a% rC2T%55e*=@7["HiES5@8w;jǗ1 G'3I(M`qFfIAQ68+0G[.3 0_!r]$Vd=EǷ O]]QRT &7*u>zr|ܼ:7 ܙ3J<c%MJx#asfц=2y:fd›}Zp/p&rPpW`i G("`NreF15%W.$D16O9WoE` =)8w-1}*-|*`"DRYHP9 uW* /2&QmQQu,Q Vv6l` J#Vⓐ{B|z+ 26\哝w2FH⽔r+:A = xU1H_7xqC]=T9)8J<r<ͬ缙촠s>>&Y%33jxy dny1:j?.V=|ѹU ߔaĘԸO̸w& |b-=*aSUox)8hMwfc-5XJy7v$Uj`EbZ.c 1Z;,Qp",yHHDq>A͞Iqǁk |},cJOy}j ':fF)'n) F*V"MCi.x XFWȈ=WMEIO>ᖺ: Ǘ1;h/>w{mlVEXK]DRgۀlyN`6]I_ʖ5&z)eLr)d#YRlWW< V@ +,BKM$~7/)?fIUdYX"͖7vvu]~7ū&_no.J6uIM95l/翝y5)*&gd~\yM#_ H,/2e>׋l&r7NWx►zeL("2Z-=2""C`Bfx}sΗs `^@m1\* =X^vФ `YѱI$$r]A@2eD # G&K5}cbl1ţJ|jצze)u{;Y[ӾB.l {{1/J=J[Ov8yЖuL`~eZV=agX٤FV + e:q)lep#"9Q`UOmZ%xz:AmI kɬK)S\fY$;: ~]Bζ֟5@>.b8*+WT mt³ǰ9/Nb>\ KP$!l1br}<](p$8ġxĐc SW4":vVŏ/n<;yh;&}>wƄɉ*JbCl i#RR%A!>EJfX֝jZrX-Zɼ U3_; _^;n!Ȁ{2xkV:껺2Sϥ8|`ݕ?[v|wJ9kv~juW;6rŲF&Ga.d\1Q]pN.c)l5r`ovgJ]`bpPQY-=@߄2 apn-iM|Y>+dݩR%NZy;1K (i^Vwl` (Yуjpw󎗻j;/c̉81uem|wm)˘@$XVkCÁD Y 1\ȼ+Zl@F!Ļ`uGp^RŖÛ6iVw%_WPZ9JP0G[X;˘ 0hVmL)(etYcp+A}xL="HRG7q?}=Xt/c' BHW]\EooWe誣"/`Y:jApxì^ä3 b757!Dzޚ&D?tOZ(FGτV(Yz8DL!;RV LdԵb):N :Qhw 8~ٶڟwE۲̇ vڶa#|ESWtr.%TTjd@p-D { 6f[ כAb1.c cVQUj|j<5[AtV@X{:sO?!k:bRD>RW,q,ZY԰mwBU*bi#6R{iQ 'V$ܘςGgtՋ0(֢Vn:)ɖpDNh ;!p+!>I83rqJYň(&&W]*䰶*iLȳ^!>W5. 3#Qݳ fЇElz;{5eۍbDȺ:Cz˰5$[n J3x6-N Y忏7v fۥmjiw|7`h:pީǍoS\`$G<:c]_΁@ V8H JJ4R(%?4ʭ@!C(0o̦, txEV jqI6D :; #}U l(*a*ưpÖ%<b%P =1Kg]a~DԖ:UPWWQ ~~!bm0zi ^&8ʒպxv I.1߈ juur WIO4d>0;+ {n5ʸ|xѺ[0ԿV%q9MZ2Z*RkϛQ{/i2&gƀh&q"Os): DiL]?h?y,uLJ+nA;ܒZyAJTt\kXY{o;qnڎҐl)0b$ڬ2+4;wGPDs1hyAЈ7uWjXYmYF)5d$6pͧoӘgpd=Q \'sO0Ŵ91y:G0#5^v( Dpq)vW$RhHϩ׫BOn[RZm;RȄLiQ"E Ɖ,[m;@' uUhl+,)M7.ʹϫ`dzj6˘lF*M KYˣG: dB|WPBP3NAbdE,'O9w7wyk2K]F$2l=00غoA9mHM+̾~whQZuq|'eUMޅ%_]Yu? b *SA%Ko"K-G{Q:n!(!p$7!&(ŨP\0*6aG#MRM%R婥ǫ|" [Xv~Ev:x[{t_ޮk+!(K7b7$SQrF?ox!7ńhGI&]4hO$6_. '03^"(-8d|HHKIyw 'wU/3DjcϜR071.?k oaR?z?ߠ;k1AOY[gISǑԆ!G6sFNiB2bB|zBNP*j"%YYB#;)yMv&"#RIFsqR;F4 Ϫ~O#`^AFʻkN`5u,wұ#oR` J#V7{UsXµԝn -!P .c G'w c,}231/_r#cS\߲o*f+㊼ɞ.aÁ_`ӗdB6EwPmد:23H= 2;[q"ĥ-AǬ_ǎ'Ur_'X#aOMY;U%K7?F0J0֝?_,.fz@7>w6N-R{5ߠӼ./΀jΪ37%^"; @Uxßu|K(8TiB8ŜI-Z?2 oGߜOIKؼ(X XЬ4AyD{mm"ʓY›3(G>PeOh\Z,'FQ1+K?+~p{yd`In4"߯z-Y6E$06cXU+64 y3j.r*>m-<}:/#Tο5?*ʰQsBq2S'!<;85,*<a$!2<B:tmh9kz&+l1'l`@K_03wN h)ӸA=XpxB@5=Yx> 0j2J %GMu@Ggh:;n(u:"t-{44 ׶w6("(T1 (m-OWyWj1ԝ-0!l* @Vp* 7b&e:B,׶)W jΧ:EB0N<y, "JI<^O_.nw%QqS]!Y88ZZ9kƆ`gq Mv&ͻ iEFK&SZNF >Q 4{s왤!pIJR1+ QҔN]pQvSd C/gJa2xz٤}MBt5=laHv'0z?],ڬI#ohP*U,Dπ PN],!I7va0`OޡPטHmmD"p(P  \ .@8ps UTOe SHTZ&k +(edC08tpw22Ix`8]PX3f UQT~'kp?wp&RQma&%= rlhRaQB29w6ָA=do ]!?0$d80 .':T(>} f5Ćې*̕!6~?-VvORTQ.9:7*E^իOi10kWV͓),pS)ܽ]q@:ҹLq2xW:R+0p01vR3kCs(5TX[4&^:}VI7>'s;^[ivMqMY{l% һrqd$.xFip h`V[޹gq2v3Wxz@aX2$U㩍6Tdj]dP5^j,`p|OfWe:0[} "|*:'Q m*6=)`oH+#tO\|,' ǩ7!l 4w5T')Wg!):|P'Xt9 eҬ20ZLXIku鹡Qs:8sk~6MC3[KBn;Uc hgԣ$reQ.-3Z9%~o`uI_Iy( Pe[P[i05HLVByCGh b@TD^i=lm",CM?Cx\ɤO@mܥC>C0Bz#)_ nm2ڷөSdŔȝ=zDڡks;wC|Fap ,:??B08F!&;! w cpfKIQb8de-XH,~zӓ&ޮ'Mé֯qB\f˜h eJ,<bsA䏋^ۨ#4 #Yx$)f@QgIJ& ގO]z9ʸw~5XQQIWzv*ph65Lj5?.~Fap2$puON+fC1$*u?I;B(8 9x#$P$V#1?[ v^K`ZD ^֤vOh( p$Q 9}Bf|CHuF0<K>T耤*IuXf0I Y¼W:KwS1H#"B&_1מTP`RpaV"wZiw RQ#lLR oUh/يX*qNs42FQE055l?LLzX < sy{lz)vۣ [n%^?~|ϭ,VʈJzV[=R Rlxɱ맷-~Ky85}v`<@)r ~2^=>wgؠЉG`28{Wr7 <~; 2~IId6"\O_ -57 ´~ŦbM!Ҽ}cǓѽ]#(~:ǹ£f^F0dWT.ͫ><]}ͽeZ'oD-3֦! V)5IwUUM@FsN=XPގgv` ^=Fa8Fq1ȇѩF?v9☏~u 5D"k۪9Z-Ѹ> o;s`^וk'3m`Pjkt EFi1,I! 4,v:;N%ޟ;*mX:zޯۅ!Z@+ xr9NU ~pV0ÿ$ *N!;N%twJ:TWZQ_>L_3c$CJ= 5tLzZa|\DEotH]88>}\ip7P}kRH}O{rH34<b]x9mt#>r=iwI0X,` #_;ƕ/vXqqߍ+UlGӃV~Ķ=RVG?uX]dc=o ,woFA5 54Rhy׋Qh WfќYr3x?^EVOFS0A08~%6(9]C?Wk꘳{J<ݩ8lI-x1-Hm }÷u;˫Q߁9^& ZCGicF)7/F|=買l3ROʶ$O1dH?)Y%2#buk$,:cC( ^[1TUAjC]Kfz ^=7SXƱ[^xq*q& 慛M ys%xYϵyAvؠJT*h8/Q['ӢQMkSEܐL~ڂX»~h8&_k?Z "KZO: 9OgFFt ц"rٸlX$nOBkRhGf\f^xZ?|K7.fأ[>⸣'{qyAp4֍x>|Pۄs&N x/.CmWv}}tT%vw}Oя찿9 V:Z h@?w>R<1qcde7Er#Y9q2DYuفmm !+*^;?aV?9mnMglB1Ru̶b|N(K(He8yR*(Іe&xHv ;5.X6X/;F.;oz Xiy":pI`'aS$aaѧ$,UHj曪wGid+MSU0 ޾s(JпP$_o^/\|ŷ_]\]lßQ F1g"\_nXᛯ@oN  Edz6cH((f lFI fHR h0 dd0kEsX(Y=RC~bZROcA`bqcu6RI^()OTVSNTqF JЗc4B1L#3P;(SFSB) [pMmVu]:0l|Zet|5.h6z; >V/_vB6~6511g7Υ_e|Yڙۯcgku"n񫗇ުGшB!l}nυ VyJ ec ъp\ T4&ٻ6neWzӕ~R'= дE5%7֭,$=}HkY+# ]Ԉ+Cr8<{'v@ACYB5 )<BONae0 [ JvJm`(`*(:q |hap &#㋤R73±X!*<En/Ep$ 6n݃W-a3Ih(Dx- Fb.sc7_KB`~Xivmu5`]?/׳ udOpCҧ/k33g–Bp {jc΃L9,N1Tk%i"kf՛o:ֈ"]{PTS\ tՋ OGHY@>DY# "A޾8''bJ#z HŅ*|5b͇X!|5b͇X!|b͇X!|80d~)V[bXm!V[bXm!V[-Y/hŒdD,KFģX2"%#bF-Q,cɈX2"%#bɈX2"%#bɈX2oŤ'88Hr.2isl9AI䈱^"%!hLK}n9H[$#n!'4 1 HDG=#$$i@b$ HE<(8*(5J#ʊ5JcX4";@Z/5J#5JcX4(5JcXsz;<5J#5JcX4(5JcX_;=>aL@}:ɀb:)p X\D?gjL fBLJPJnX?Tp4s64iȼjCX]fWJf2|;LJ]1y6lhVtd@y.Bz"?}SʅPT`bmS4J5^R Cm1φ_&r<'yVv9:M~?*f^L_.,JnɏSleʋ8JqYŬ $N'y8xO!1,*;sJk17H 5Lq7g`7kQnpmg]ٻdيc;q`ͯ^&&7-pzhwO^VYÆג]D2y%Gy3Schltn#4ܮVu۫]ղKw9펺6͆k=u tU2`jsَe"MQ紥B\NQvuUyD/S~ v-U⮴U:6㎞>cƂJ43d49;.2&S*i')/wA$sږ.$-ClHR\+JށW/2Yz]ׇfP2Z!Bbxs &̥]z*VgŃ𱦒oc 닏wm?zf]Q+RKZ`&jInΦӐa{mMP;զ nkW©L8촔Q_YجGxj->0ާy[҇"]Ɋ(8+ƣ]3/\UfK'\BeipH08^e 2$K۳tiCRMARtGF7۞)gd" r+rЌ\5ƾ09nDMl%3!CDW,m2.W7n<;<*AzِꁐTp%9Y}{oCUr)EW Ws3qvab3ϗgņzU_q:D'E=unjC-؞':ohB~6/g" T 6>0jhuTR8l&΂1\x"|I}1IAY?M j6q_C 9ej[ ! &0w4H^yda9" *g,h tnEi\]e|CV#Wpʝ68i:Y r˕x$ oYgſ ^Lh2Rx0$Qb*(3;Y?j!lBPK GU߇5}+/S l*7 C }K:ka݀R뀧 }r]n5˾\B)DqT dv}z "҈x\"*!?ĽӖ]nP=BQac :,Dvd|;]Zx,Y+m.0`O*5e`;B&{_y'0 MAb5N!B")!]ZxJy Sys,5L!IN| >M|(CXm ,vj2FK0Ff;B/ ;!t F8R ڡ 7cXlYʴ|jcL!]ZxzyțTL%:5(6E%wielOSV/] ++SC ;b7|,eJG'J!We C u.-ͬ#F3jH&GA| ^Szόq@z=dBg&^R$A)*@HU%YC35 )<BƤ4 T) (퀱  )bpǂqI@V䰞M%򡅵 kJ5&#qN^< RςԖnr!HOJ @fA-2 ϩӜIm k3ęɽ\Z<|yNx2sE8?S㼫f+YgKEi8\á7(6'hG˻T&?M'>L'Es3_I~!-/n=]uCڻ)7vŪHL ͂N뉞]ۈ+7J^k"zmU u*8t́}Ί)QCbp2-ml&`l)E&>"[+4h22+ tv kΪ&Q9*J*Sߔ$&$]|^<O}'gN^<=5ˢQ:}~9h9_ݢkW_5(.]>&CIrKnJ#~<`INb7ڍƾO 5FPCG/昕J+'_e^43Zá4h ``!keVp4n PY`+2kt3 Q쀴oyh1RTf)5‰rYoVUW߶\D鶤Ul ﶬ<]#m2!r@qs&8Mr; M1ٻ޸n$W"/`6 / >c#iIv=ݒZZJM[vЁ#ݷUdTUUqNWjcL䎮1\,1m߿8WܘnzN?'eӵOyFyw|f;6ymbJV#+o֧S ST[ht<rsV/({?40L&nGg<:?8m8;g:\G%4YIYt-{;pձ=e'+>?}e>K-TP]_2€+[,ᬌf- +bnJfn*.aZ }r$Z^0e .rlQ\]2YK= #}r_0`*Y[61??杓Ù *:& Fd@]ʖ9G|J -EI$}JCs{~C4zLYU@v$ +˙%#`5: {ye%D")UAg~#ךFd50NS+لbbf)d!;3 TS$d]c6wyq#ghs? RԨ{%8nɉR.[\KFᅴSQ:FE 5JBnH 6SB}#< [j=@:Q>.ةk-~jCzMVk_@k!B Y2Ù/N(@3.E*nB%%#`pj)$ KߓAU9痌0y;pmH^PNc,,Y2F0|qY0ADj6Ug%# `IG;6DE`EhP\ %#`'<~Lw:YrdzD%{1cǝ|jer#b6׬M.ǖE*L]R[#]]<IvʬsGn-u# + uǫ-S>՛/u!d)CjuӘI᳷3Q-fmy釓=) _앸ռ?la;q䣲/N-quy?Cf_xC[)W"-C': *ɠeߝwi; 楮91L t#H1)!~&epu \w[^87icH:YXl$\6nd%3dٰ7iRGwL?y\%e"+jJFvz-ɐ*IT{RQ{O% UR{7r X.sȕ Z}|}97?M4T-xp*q˶@f~_gFCm{"H4kБRO "6l{QwoFPMkYDpE6h 9?F`l@/#.>?OX쓡cXlYX2_wcFo{MR=bC<¾_ dم&ֽ.p ei\VbvG8vS>eQ+d{n7o_5G5iمH>g _ )׾\}U_ ? yiKW&|hXbgV҃ŗl9a>h- gv!}|$̿]sON.דo.+^7]_(2Y#S2u]`ڠϦ>w?<_/Vta; *.0rHt9Z#6ث|~^߾bz#:OJ%i}&hn]߿>8_Ebn,f%c0ٯjVJ~={Kݳk^zXQw.xp.R_nT3 ]&}+W|g/7.o^<=]V ]W1?mv6|5-\:Fxڡ@^rNj3;?N_pP6 lg_ VE+<{i>ycSˋ+HӋ(&zi/ZzXl&nmWh714ђـ`F]*}]7<_=v:+nG^ʷ߽sFwtgu(Mg:X2^_G s|s/iw ?~3_ѷx_Ɩ'AQ,?:Rܫ 7xRwE, ^`_tDvEsz?bZ/Sp'5?nB} Hߩxjrf1V߶SwWGߏLPb|ƷJ ȦVT)iۋ76#LQs/m׿m{ eXC>VnQֿ=*ڋoۿ諒>)Kr9Z"e)5[Wߒd hUI6h&d*&xKrA|ղSH1k-cJWŔE`!nNc%֎&44$+|'뛸q-{X :$%%B[ Ӝv"Ԭ4IHTQQJ K #b}'Д#e`[$4nEE DG9+O=0 /5`;~ *H#Z PddKJ|-3ISn X"D{IlN:d :2vC5)4.U+TR\m^g i^l/wcez8o2$ {`XkjW:^式X Il\l[yzr2 $ǩqپ"\(F `ekmȲYǻ &~؉7W[ФBRgsHIXpdvխ{ϹUJ  ѩ-C^b-rI`̵댉pȀg:ciZ뗝uKk׭,3_+fHNBTRX*r!u@0- |g?֍4:ղ \q5*h #j٦wu*PǁmZ0hƋA G\UАV%^E Hɭ -*=``>\qLq 'j z5:/Q%(x$H&rZVȼ"a>X\gEajۍU}o"A)ЭLx n`mmG Vg{ESY_~(qYQ"m dNI$u Onƻ`vjY.y6d*,hb ("F_-'B;Itm^o.-E<$ x—ѵ#De[@ @2 hcB^hv)0Vc[ 8 9@g@!t9n3VU3c-NmN;(@ "(Zcv6kAqX-F|d)ebzc"Õ,-Ƃy+ s.pYTeH! cH>K4КmN?cyڲhR1M`QIRTZ[fq/z+ۭZz+-Ia. K'y*0V't.Yu0q ٻ[O`t aigɸc$* @][4q [-d0L&Sns`*ʿw4,E4^[f IԹ<m #es7SNF-?$9{ gj  xrH|( Z5HJdp A(P`AH:P#F' @v߬Wɰ$>?(yE rr\ \1[v)" ^ѭ `\ة %p #EiJF|R5z{֫۲ZB@2=B mXi(=+NFp `Niu)Fndj;R3׬T g@qQIu/I d:USz|p-c|{ɑA6yQK:*fD:8FD o7PX˙@̠:t zAy!͟z*8o`p{&3tQEU2 8'@󋆗 .ۍ9mzِCL*+:_M/cdaVMQ.#!2l]X豈)`x-qT(QuvEC@Tv""DoaL2Հ]. q2o~mHISVЀFH5\,5Ȩ΍mBMaʢZ٭ͮ SZYz'!4&5LQ=y%t@Q Q@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJB쐔@0.y(J X2@R9R=C%eJjR@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)3}@JjRjs(Jj}JjӤzJ S"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R=_%H'Ι@\#F k}J X,)HR@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)zΥVץesqrA/ީ<[xI;t.Zp VJ.= ҕ8N0jsN pRDCIZ]YӹG?Oіyccd˛*k|)9V6=/ a,[<9v~8=/N'YnLlb("jEߏ3j/tE2 qmL-[E*Kmb3x/$ /r೼RQ4}bj1.96^m8>{ecu? n2]Wзe>x٨fȓA#թ"me0I;X,J+u|qٽCRj@HLMGUs jOU+-#$ x~;݈ľۅVr&6jf"6v_u<شFIojs"'͖gj O:-ffLU8X%OFتb=AyގivÏwvo-mG8 dT_7uѮN?mQUF%曨G֕qJD[\IW뜎JIi߼ͿQ?aj~*De.xۦnڜR#k>d/F{'>j<ҰG}$]"ye7Gԫ}e.ڝ>r裗Q#&. nͧɼIELbvo<:>ˣ%XQ}ŷU=D oC/7 QѕËzy%x{}rͥTbOٴi9Cf֗΍iq)"Dd;ݛu1P1oP_W]ձY㯔 &8"jwyl#_Txݪuvt"mmܢ/w;˭Ɂo|};Xʣy^џh\lcvyZ|w=Δ?.W37-U)D `o uðgI 9}u/YtS(cb"Yd[EZ7e|l$ٽ -z3ȼTh]} !+Uc;%f]@5CYk`Ay Z K%(\U>pUPUɇjeL2\9EkcvC7W֝.]yoO/7Ypj'L^/|qiFk9YTq$~ؘ8kk !{0qaTb,LY,/ &$V/K^]{//.\ S\2>fxqQ|12<;ÿ[!rDBXo2~%x/gcy=;]sY76 }=Ҭ=vyt:q;rYlճ{ !՝``\5WC!rڧGVR?G"g$SZTH?N&gϾ}~t%WGuYF`zeڨXG]'GȵA%bDHf!;ĢrՉ?,|h-]&Kє %Z Sm \yeuU1Ƈ4FX M2FsZ]^u%65.,1?u+P_/6>J(4bLX~NjTLذy8쿟n$8ztݿS1nF۾x­$h7Pw׏rW7׭0l6tLNm Ȩ+Z2M,21mANy2awZiRʖe+TNxaT]La-_E޿/~SfWuS!U_C_ÿG׺1Og寛_׿k_ɥݷ^&6 gDIG[=56z88gYBٷB[)Y)C;)]yϱ |nnyCAOr ]Wo;^~%p'qw7CZdSת*y(ujqOUt@Yֵ,3U|` !BQM=NFad12n/l-N|?IJf苻{fQNط컹ԗE¢pB68]oW*|6h7ZV#y[tk"snGh-{b&mEeV"zFڤ1+"2\=Cshrs{e$ݵ<…`Y*S2:J0e͞e3+͜ݬ?!r" Z2L R+Z~cv31]}t]Jܾ~֕_&otj9Tڶ[){!o$"W,k)(+]Q >TQwUI _MWy48Y5EXu@#wUPE сsQ#*.$x JZΙO'c}KtLi-/\=뵰z̯&:$9'FJ8Gq(JM $6|x]j6NcTmVmu9S ^-8+RU<+݇X_걸VOZ \fڇCb׶C0Eh}^/7/b}}3vrU9FѠLkJ)]|_OL?/pd?|.|b ~Z$dyd/Oxu`Tޕq$BKb3r߇:@,b]C""(ǯCsDr6`Y1]=]O ^2F.#4 C/0}χ\?Ê iFfco&Yu2ɦQ:Jp9>SyxW6Sr>ýaOZzQ2%./KoKoÛ0N4hPg7 nTJE銩 XTS{{|sݫ{ͻKL?/.߽}zGx.Fzo~K=x0.~bh0dhPě ۜo0.'mNaɱFy|M|GpSKvPU*sYz9_٢p$?X <|F5`@3vN8?~HL4 USWe`%1(_ 3 lg:H{Ɩ}(#EDM`JMlNojQ^|0zgT&ܖFdKjk"6# 54zXKf j rì̮ѡc U)Ʃra0gxSW Ŏ't@A QXE=SBYlAcW ^ &QXu9J*g?|gc_x51^@9ʽ[YQ$ty d-W>#ջC^%V#7ePDၱXhT V0bOSd=EF4(r%sl $ĀC{KRREKC[mi=5*C F`&CL7g)Oc.*5UٖqRkI )&m*sbc"h(Cn`d};2]v'1#dA{_fHC&wiCXt:^|'ƶ- `*\5OU|T6IޗPhO XY7mɥ.<}l9~4wR,{Tq/U&soL>Sԩcr[ i!t5]P@U`Sk:BgHuH%Ejo`&M3z8jB~yeQZzeе S@Wfu$E5@liBξ7ڄ='.vexI\(w$"=*j|=#,RFXp4ZWYy-9Ұk,"h3_̽*n|n ޑZ}_\ {u:y-y-yҭ}[-ݢcY&x9dDGEdQ0A[R.k%%RHDc­tkQ R֎SZ9b+v pZQ 6}uBhTfs$#K0,pwꮰ'i>u$%+moO_],m$.L$h6:?|zhN*nGy5H^nKG/wP<~fpW[ĥ世p;i9>%NJqbPma U?z1μ99y=2Gȼ#zt+iLoTl )zd^|4zd^!J+&U8*zd^2Gȼi0(zdԕy=2Gȼ#zd^ ,zd\y=2Gȼ#z<a ;x؈ЃqB;Sp F$+bYJ f NvMmƦnl7Eǒo9kK"GXhquiz?fo5;Y #E$ȴ,"3{BHaILy=[0v ؒ$}Wi;_a5K=gC$pK}{.\42ׄ"(0G,fsJ1 L0ܩ%!u~൵/(!:%#ށ' H:03,*ȁvXŽ*'<zW$v? 0qiSڬ؛u)Kx.`1sŝRC.T1\`8 jQ^^s&U4=$";2&/Kko]p93suh|=/g+.e(ꝷ>__b8 "J {7kJL?u'7~|_/`ud5D+Χ#1|aH07lw`X.&-;\lM~8kգN&4j\@QG).g};. >jo{)Zް .(tsc__ oÛTvQ+St4I}'5+=Gû?%J><SRTo\ 6ɏ}w7W~y՛wˋ^\{ ԏ\K#"E}/>mE[CxAo2ns9q&' KWF|pt6\őscn/C9bT`9fJ(:Q9Μ0a ty %n H0 IjʀLu(*3 l>64-4yQF=_.  (MlA_^)&*6*VzEŏ|5 YMD= Jq%3y@HIcaVf1* TK*U[Mΰ(+̧>5WE};kvc* ms0VQDϔP;u땂W9c&QXu9J*g?|K /& hB== G[W~+c:jθZ|`VgB%*7M(X0TW媥T- !9bw-T1  U0E^*FDoAsK3qY71j rBxr? 9*L’v3"'ڃSxqJ7Dd%0l\ۥۭ&NVl# xHsU!Aiʹ-zD*^#DctD V 8GGF=F^I^GnґӼy[ټcvJ{wx6ee_ȚVV:1bGɿ@­&Bu JWtAi)ҧ%IIr+ZN919191919191 љ$#ddLI22IF&$$#WgL&$$#ddLI22IF&$$#ddLI22IF&$9I22IF&UKZ;Q`b{q4c̱:PͤY[DH(򅖜N)b8 1fx>4WPs]^Zt^[cR;ki2~sVk/;^s;"*{U٫zIvn>N53:WҔfHu$HEB p_|QN`RGTPJm6>鴍dQOG-]HouV?2ZV/ @cnl3:bkhH 6hx&( 䥢"^)EǸ.>?ŧP)np0sH[]D‰uXTF]O u=\X\L XVQI U <&՞{:FY^1Tk%ik'&, [̺^XjeZ!''^q\ Zg=^ݷ%puAb pSUF܏ yzj5gXڣ#{AN9By .hEN(BFp8X0֡-VxPq2!åh\:6hUD&h%dWDhL$xX[8xH)wWJw i-!"!4aZ{-6E :pbg y K;^{]?mMj.p3աI`%:/;%/IUn 9E_B-c+8tr* B /lPqx\Ñ%ҠȅαaG cD{Kk/874D˹E&avwd֟@9>XQenTOp&cqq9VCUqg.+EW9ct\љ'غGG؛7\-yK<=CsD{@zw&أ0.0z9gp.'b}퐚fJ 5+j:I+٩$YM?C5UZ!uG3*IWU| IIs9+PC ;$쌺Jjy LeuԕԌs!uPFWӮ+SWWIJcvQ])FU/M?HFOZQMSo}3V~w@ 4n,ھ"h!I\$0םQI\̓9H*sN++ u펫H)޵$"aha %ٝ; 6y,_5IIDڔ(ђ ƉE_uWw!|vWO]醢חa.eo+~lt?ڎC\sWq) ]-ܕ>]EOU5A'2v+-c<3tW * +`qW'G̡N^mS_<Ն̒mSձ3*{cP{B@$c4d}_ntKUF/.VYaoz"[1-Jj _e*c8ōf /xڊe]:ki:onhgC8 %kI>!@$|6sltEd95hoY Pn,mw̧6RR=iWz~ s;!VBߧk\:%Qh|m=)]% Z6ǻZzLo5m;CU#x֬c0ŻYQ5̥r^HσVv/w1RщŧA?j7N0;}cx.őBrSJ Y*r*|-G<|gObI*+m /}2\ o*Z "|tiGookVsLE\0P&?ct0)a޻d8R@`ПLKWh5moE}k]rjbu2ukWs&/e ᵜօw/z8NgE\(n6Cc7<RM?;d>B|ZFDc]¿KG~h(o?WX,ʑ_b~-6f"_[ G4(9 lt"+R%˒fJce;i`BIUf]75͚\o.n*$[5*7H?~t0 Qh'toxHg- _ũ4=3`~Og[92K!J--oIT&_!WSxJ<.oO+켨l߆95F*X(U6U330CV75z*gcTF6Fgg\ ݸuFs7BE<_X"Hv)fwg/η%s(oſ!%g}( "a0@&&K=zZK[7,au)/MpRVw٢JfO4rl0~zYԾ4G\]_ƳuˋIszۛ%ٮxa׽+V\O|&:3*AP v?y5,YzC J(c(S6$0JPCCu,9|v3,b Ƹ!x$Odџ,xysNf ^)_ϒ4&=ϒigJ/܅,*VدV)zh| 6)e0$x , E'+dT񩙍';s? /wOw&XKL% R[+;;E0?tҶ-g}y*ȀdPm2a<7ձ i'}vm~erR_/DA7+oeu'ehv:M0Z2zχJgexp:[+{_8+ernI~| zVFꦽ;͠YOi/K׬O 0kTRKs+m k /qMp'XdA xl"JB!;`N{!xkWv}K(üs@dǣ>":lc[5X-P\<+33)'IiFYyZI0Ţ2@P!EDb̔՘U"R9LI8F# B*ܤuk S{dƱȲZ4`6ȵfg*c*)!&1,d x%X po scP]~?q|lcmR` ōRZ,fc)9(5(xLq(&;k{WSSC7|3S}#,\(b"[LaP1+!8V iφau7pː U 8<)*&(0S9:R0fM;}ٜyaO͈r_u83/| -?!{r6->xU8o5z  _I<W%O9%il-پ +K^0j(<ՔudH=k`$M{%i?0~M@O4qVK;YGZ%Չ_Jv < P?KQ{iW|Q ¸f)}qeb_'Ɵ*XˤEǾw^۳`pqm)ȸm+JM5b?Lt F/yͫ??߽_~ͻՇw??e Ztk~nk3-H}zsu1 r.]?MtO!k}/["v($? J>zAv%0˝GY/濙?:}#v_݉!&|b oCp7rS~7$3k.Z@mPO[qJ#E\C)GbmYJ5LYu+YRyuިhu1z]k|Шt74ҡ"d"p" VƖ:|..%Һ,Ш5d:yxm]V!rr0V-SBEDv…adO9c XD,1>oe+g⦳QK`vyao:/.7hn5;fCF@W#T#*uzYKV.OWRP*? ~bΨQPp@`D$inIyZs?%\BqfJUL}ߔ8 s ǚ8qBX!(@|@[K#c^a I#X[J0N! caaΜ5!á!{I{ca>qWKG]5XqYR؅s..la9WNq=B4 Q8lx$|cCZ/iuL bJs\ q6޽oގv[xվ[gg&L%Ocs|H`TrM:&NM֧*i]N/P-NS_p0JalH;8~LHYj 3VsH֑0Rg\jWP1U$֕j¤<\o4[_'0#X-^z߆Aμ |w}!Nl~M>%f^ozmoJQ}0ryz_\nD6'3z$8.7ڥrkդdu=)ԵX'ӥSZ(k Ɔ~M!AĄeցM1,V1Tk%ilXk?XkS̚/m`;Խm t5lcwŰ] T.i`Q?zmō6Hpkm:ndEW5K2#d8 GHaF 3*UR.IIvӪ@s xL:ǩ͝cgAL,b4ДЀ 8ǡ }ŸKiB@SI@u`/>iϥ#BC8/FpWLn:|GJjaeģ f4N] .x虬}Rl5vATW犂 At#WMi9A:EP6%龗s>IJ0^lh":|jM<54Ne,]i<ܷjDI:\].U|Ym,FໆߒfFoSqyRu ~ɡĩHC LRM 0<ǃgyyII蠆\F#BjRtJHuu[[`!™e'm. =EXssikhޚ#y[8iIq1Oh[Q FAzO3Dt<y{߸Ym}d,r,Z>iy(},Oԁ2BIF='uWGdn"DϕZKuy&gnʆOyH\2K2xy ҏ{6e'm <̽u].=Su ~0dNF*=":b5F)( w2<$a ATc믊swVBl./.Eʋ'?w8>|o2޳6n$W`07;$8 A/u#K>Qw~դ$S(Q6i[N!U (68|0*~YxpsR?{鹸+I?0YEQ@}\xraVg7hb?guw*b?XLUʢlmz5m[8w fIsi辟e~T}/o>WU+nYjS|3_&e :>|EMY V8csqMPc9,)g?~l+%gK-zcyjfuҖ*:՝7G{fּgS։eI܍)k>1i{Q=V'5kϹ,HccK F֒ء@(x]a0„XYBd w3lzP1rH 6^Ԗi4f_ͼV,`]5 |IT~ûl6M=lca1|3B/s)#Frް?uL2M0\r G֥NP82\YbUr&{ʙ+ggI4Dp2 cD{K|A4s !MU5*Cz\P Vz1W(5XcUg=Oo9,+?g*g>=.W!eU2fi-V*&=wM{]P̓cw_HLc5tGP $v_sc̱|>!$i/96/H 9Zr;O0Ó!.%#ɽDnݎWbR]KrVRN_[9ie Hx6{]&f%T[a*odT^ESjt{G*8ǾmӸ*L9-l>ހ?ϲͯ>nږ^'_'IsbW tC&v PxxK*o 1e2Ū`BRI7!iK=&2^TIz! ޿G. K4O$@r|)HkV $&r¥ EFk!x0k5f,`R ,0-<650ETU-\)k`OURbx^mVTUڻF֐P7w o+0(N.j=W0ccJ:rƲ~U9 ¯f7ح"Vlz%|W=}sb;8_lTI1})/ӗbREK1}SRL_7J1})/ӗbRLUjB),97"RXK`I:ZJGXk~(7V;WjU".C-E*B3AМ~u["\Tk ݌ Tϫ"p1"]-@N]8ZSY]o_¤rU,7P}c| "G!T`\}tAeT%,x0=Ճj`:wk)˔X#[Oz$YL*&zZ#I=zk/'Mz(M?#K$'ʓ1)re tʝR3[]ӼqWIXӮk_ߚ,nvV4:Ep6HSfBbly d01g/+yA<b4הМ 8;{}̸ZLhtz{ڢ {RcS^ q8/4y;@>Nf3zrnb<7N͈+Kk8:8@yo=7]ZxЧvƫݜqgDr09|YyW87v<*h x7*堞@>_LF5|[OSks?N2͵R#d?sA H";#ti65b0j}Պyet@ߨ$Vt&~KB x=*uW<[ٌ3e*٪W\9l㑏6*9KR6۩#Ӂ8*ܶ$`pS(Tx2L&zb|BnO:DSt![2X ϶7_hy@ &#{yR3.<6_فoz|I-R`mbP)IR9Pgއr2zb1b Q߾)K@X2tC< g1).+{]r@3&}9Z߇nƇPRV 98%߁rzFyʫ0픢#1K-./ 5Dzx[ w| g3n<)\ȍi&W˥y9OF3b1 OxN񇧍adFٰa>jdQY%Tu'YM~Ϫ8e]kݲ81<³ f6Tݮ˦ F-ΖwnU+sp?S}G/o>W&UZj|3_&e]:>|:Ee1yBm ^-?Pcv7]CMgQXV;=an`&}Y=Y|?G{ɱN7W|%gS`Ɔv5{HVu2av6ԍVvu2.41*?e>N|[ `6q_CC%>F_FGciiY6h~-&gIjZ 3Mk#,AǢ%&})vϚ67?VBDFōK.&Q+Gye(`k U cJ#66HF/g)OcPknz>`>E\8p0M} SY}[&?b<޹?]:馋l6d㩋eWa͡]<nM(e_}b8^'_'Is1nbTEg;jBgeͩpVh&o pL=Ad>U#:w|i ĢbLU^l'Κ}`[!>c|KoXd\#f>[r-:0*]K yGec@JXQw¬a ]R}6lgS#-Xgpt5v_ #Gh~wy_ڍPV .0(.7&RKF/1Ǩ [,Qd;G( oF0 f6*+ 0jyVyCߘKyHdKQ@F}.Sξ,S}VCQJe!$tHّ_5@}C\};*2V`}& -8X.t駓E~J?gSZOy 0CMoD̪ثU=)gPܠj$_D+W n&-z:VS{7aMduP`B p4Yu2LwZ {- L MaFs+%h-^gLC`]s1uz5EA{iK?Xs,9&uWu h2͐.BQ8@nT)YH>;imvWٴwv!7=Xy]@.]wTxM(Yx_X7S\B[M( ,3SkopHՑ#s/vi;d{/t@܊AnxTPR-Xh XXElpbCH-GMX"b<C@X\(n-& ,lJ % i]{ȅԁ*>)*P(שלImM{DDbgbSX*WUzw=˙+Z2!y`RK6 gs\@gJp1/8_`oQdg81_^]ٿ?r;. \\䑬/BkFe•#1zuÐÔ-Vlwa+V1qAmУ٘ݛx8[;*AGLrݨk:JpY8Wi`ХO bxGl ]|/_WݑKG.&y _ܤSLꪛDB h0)tG$5ꏻPQa0r~y;sO@$߽Wo^~o.xuy?|;ؗqb-6Y;Cx=M MFкYW ø495.ӟ&zOVKK!+<(`8:nZ]C`c YɴrZ3Nw8ItW(v3!uH0 <2i9Vxm\ \,iQNLw s<Ȓ]Geݱ]aSn/H"gǶYӸ^\]%n\WI4~bzjo8Yzҏ6 ϯ[Õi$X`T# beD/5eDDL@>FQ4`pHPuces'mY|OO/4{Z}N@ #nZ9Jٛt|K U<=]g[͟|Ek y5㸭Z| ,2WPfK-a0,B %G&)Ϙ& \^LkPzWՄItPM]@x0=lH6MWG1[.wL\nrUt^M]&؏#)2Wn[X#V8pJ|̟o'sOo:<m3> +[-3 RUyf:ӏ3~7⭾BbF%Vv=m@q\ X|޸Ӭ4'Mh'Pk> u;,xfphQaې3˧xBiwhV=ua%3ߍ$n#hq SEkX5W}XMb˒.WWa4QܟZ(ڜ-rm|5;˜l}Ŋ;v%"MzH@ls)Fp:@߀1T)b+t@HS 9HysQ 9@>c泮B LӐ}Jo=_x }l %$EiY6,r붾ܛ/5>*nmF遴9=p: 1>V^o/m_sZo[B]ajZҴ0VL5qW D^d#vuP{ "UAJۓjRՙ  7 &a:]}[Ƣc\Wj@W1=Ŝ9s˭#2)OKWX'&63SMێkΛTbB! ;jg>MLj9+j$bPgݬI~<}kE:[f/?Zyt *Tڳzc5mu<-{tݛp3罫g]kF(9Z;Mr<"W&kZ-q$ZkQXII˕_y#X R쌼7)}cPfq`,|EU!`}_?q;:01VYE 62y6`#*$;A At4+ỷ#p0uM5Nu%kܠf#1ʈYc}fHD,ٗV N@bS?:Xhq%ї cu4REbE7r$2\2 4 "T(+w1 ;#D*u #?l|:?h.X/UgWwNV*bC,u Wwb8k^dΛDj0Evo ŪQH/?_g#/•7f6ZG~;0{_A֎nT=ӫzr_h'P=(f@p[\P.^][g<2Ba½ 5-hDIr½Uzt0]lx-y,OT7cyu헅旕Tw~-&5neJՁ?U oY| Ϧ Y/2)$~75ҭ4k3^SµnC. rΓΓΓumO T/Px>λI9Tm6ȹQjEw9lVJ%>g%OF,o&dž$Ѥyz>~7 m_pWh=<P<2_tp~9p<' +ᣱ<0o9n0@2ET03BSsz?$6GKP2޲gȦMmO] e'îL J곫\uˮbt&m=a^<)9e!21" Ҡ(K9g). Du@cl^¢].ΣP핬BZ[;Th VhD8baց8:rYLc$TZBMhn Vd~Qw1ˢj ) sa%[¼Qs%7KDFy.c!=CED=EHX! g Np͍prF Kigk.*}Z7 ΢ &$[i._ȹ̥rs~5M_֋\`;mP& &ARLڴ:%5Ky@;ԪnfKƘ4gi&n>蜟S!ЃaGP -h=`4qb1Xy& iE\,KI&YebMTzk3=\YΝ.7^tr\ʭU) lmaQ猣kL3*7`A=c9zM(IK\Y)9X;,wJP4aJӨ;^RO.Hӭ6D.4m5Ɨm;$;3kDiRHҔSS{;M% sFJ`sÜ9Q [cI"LrV@22G2qNX䐳"(V*&~as:By:;g#JcTXk2ȁeZBe2ƉW7{usjj}>KhXZ-[0Deq7 rKkzw6}t\5x3h)9qƷO[nwOꟹyհWn SqO@3T,,ng`CHFr/p}T6d!hIZ'iRҞ@(f5\*jW,B]W Wedv'{:GǗ3ʶp3hPABH Cl(%Pf wG/z‘>תk]{jgPvMNyYs"W6 W.Cν^[$\l]`sj{dM6]A߯ͲClY`ienWMw^ClB0V7|otvKU)AوuG ϋ w[@g bba=Ʉ4{_c=JK95koyK]M'7)7; 0=X2흀, ~_^ODGg{+E [kLQX4aN.B.K{$N0'Ku' rug≅beb+ooL ,ƶuD=hKmlNU6'ݶ ent !ixbw@dX V,78'(rBx)v떁vS?"vY71I5trVmp`ެ3l7޺ǜ3ؤ}DmŃy]SiaDatcŜ*h3χ(/Ɗt'ޖ'-F_\ne=8nd{]25Ӗy>sg[(e+$(Q䔧L!߈堸:pOuh(ð]. iy֥x*LdO,I\ļ5hđGa8:V%*LT=wqOdemXop}]JΉ 2Nhu QrNP’y 9P亽qݥos fsiLkTAhHI*H) 2Pp(h,ZWc83Bӡ/q]pasDJN#HTyЍ'I^&N/xx㙢=R VJe."E0L7 S* NZVe~ H ړZePg@G9V my Єi\=nuIY>Larƽ|&?F.B> J}s,C)V/?~pqsa-L؆]/aίۏV-:huY Mz4ĩnIGl5#d0T͇׽Q?8y _=ٿ_z9)Gedӗl]"Q'9 e\yx>C}(r򈾥QS6뇙pr͗>^& YMd&^~s4_V8IozꏻTqUr+ǽп:G!op??|wǏ?ӓ89;Sum# vk~݊?O~zЪqШ7ڶo0dmrøg]ГF@5+#Yp|>\qNqcdAGφUJ+[YAƈ-;>,cNk77RaBg-2-^,sR0dqJTs;Glioc;Oÿ- 3e%d 'e{2_5z3eMba^Jv0a`^kd]P*e$49)-K1p9+ ) D:EDIRT(cɂ|4J!AC8` sY˭YwgC~JU igp]|q{%MB.7\cTXrlxHo=nZ¡h'&sJ>K,h9@୍c!TG`ͳYM[_}ߙCo^mA,u`|5[g"Q᷽o$Z}mZjm~^ G aL}a&?4c}c19wux_ [nk*>-l=l;Kk?ݕE0poa"!mq/Y VL t*R!I`^CWeY]w\s$e.4D]%c"tJTģP%E4ZZw9YͥP'cCKm>5?/n}~(]=m;|g3t!64p=2;FW7R_OI]xt/՛*7Go2Rd]}nA/#Ӫi2}3QTdV -^xs!1:s7(7sxe<<5V1.IRV9W$ъOmr3*@e28 ^.%CH^'mmT%20$1d5cZiS[#g2j▆>dSGUWhOu⋾|Ṅ ٺI'Q+Ĭp:f*Xqvf)U pUor"6+]YcFn߇?og^%s]ebt ޔ?߽b{ɻ7?n&Tdr4G'b܌a;r ﷣v!_h'NU2S7 a5EI")ˣ<Tᥜ-rfw[FW7k̛y~=R7gIhv0mٴufCq.uCuzT~]~٫:??<90]駽`s}3N*{N~FnzmY 㣷=W܍g^Gmh^]a,?(%$ gשI"SAd+!9e]6wow򘁶:OL; 3%%uĈDݶP# $g\KrIfx$=ZG:{63^DV^@>_OBQڨTRԺ$d z9kwU-aN"DKrgJ_L7 6 ܗI$e)gz`CΊ`(<4 1S(QamjIÛBezdWZy0vܫSKV}=jp,ȱ4_6ρTNINk?kb5OcyeWM$fY Ί)x)} g-<SYfF wx{rj)xy?'s=ʐDUX-zg՘IDk1hFs+Evm]?#Q㮉"ԝX:CRR ר]QQ8_X}{+%R<^(ƞ4VtdE(8;eيRC?Dy}mth-JR>RLSNq飶ěU`̙@s uI]#һ l>HEi,o}WEr@nv?hx@KU4TYG J_B:`^em) Ӂi]IU1>HgDa iU3PSA&G[p4H &$Z Bx`#mfu\, 54hL+ : DH[XA`IrY$53aAc 5P.)ISb{8.0`Q+'<;챺Nf 7`5Zms\IӚ.t_mvsmk냩$omndWA-Lj%jWL]IO3~]V~k2|4I[ s]O鯘t4f)nYL#s9t?}0xɫ?k<1r9>% ?]^mZhSYMbboMuq1p⭗TC׭7a}ϱs뮟PeW]C Faƌ>D\aw?h?aېџuu L΃`9#V,%A큺#"̞8KsYaA~)LA&r7(^_pHW +Q,n-C/e/! 2:`.&. 7?^u,{9JU/hՅOpQ+N s@t{.M 6j֍YSuKIx F]LK엻?uϭ~~&B&i/uJl"!qg~ ۻ) ٰBtɫ ԏ+:Ǚk^??\y7營.xۋ|D]o> Tq!zk~ n`θ5+2n.JSn(Q {A7K624-]}cF/Hۖv]/I7 ׮aU,qw[pH'D p4(h"$iPc-1;|DJ 1cʬ g8.p s69 (gJ(띍:`JSX*IEqsSUN? nD߅᫛jÉ)Z۹0y=4FQ\<Ƨ;֫x[aE-격T~N LO?۬J* lA/;-0 g% G:IKX!Z)vϻ kBغ16qEp>E\ɟ܊ ~31!X04~S+sE:X64{MN{LI2 4 |`/HgE/_I{WeUԿ؁L-<ޔTr'XrRVd:oke_{wUٔ=4}"g4` ,TSlj.0"~ 7n 858Q3eH#+05$!.c{j#XHp+4D4^J (nbi1g;Jaظuj#xb{\n۹_Թχ@w kĜg9 wS\־YJ$`ǭA>:ꕊ{')E!p)DCoenqQMq 3W)-v^{5kSe(AWj}WCl):wv3kM>ךFƬJ&rG|(ZP yd lϜ[MAI-j8Gw9qݽLd)ָo=xt|.ɊgmO?:^tzv|QK*Ryd!RlP㥦 hH0<)ty=rCirl:aj{nliCjU >пʜC a6y8œ ]@Wg&(LK+ȼ`(pyZ0-@uWb'U@m$RV[v5Y%=>*v7\܌ tj>;bwE4o_Y_Хx v B%>&W:]sen{jAz lZ x1 1K4:$60bcܩWU :Fz5$X -<Ɓl~s;qRoq% a`ʁ>2#hLF8QFQ5MGϫqҌ2$^7*Brj_9p.f^A)^9;-n/U0} f-=}5:yOޚ[no,A;}W`^ \w?o-Qbc1 ( di 30 <( տ2oU$3<(bAR6g(:xl)*7y-g#YR}>%y|T|UJd'f jiV֟I%=z~|awUݰ: VGw<oݻԔaJU+'~sDY 2-U(gaHZ(ڜkcU#&Mh"tW}7NᘋHo=&ۢtԔ)dlJa^iG*Y@SJxƮEv&<#m4{ypq)K)6d+} "3_PoY6m wg .:l>v;/a}^uh[CMgە9¯ ۲<Iv*em K"v_wn-C,(; ~z!QU1W2:=78Ŝ9s˭#:F M:"֓Wyv*8ʝ7ؓB4A6j|sWLyZI}֣>ĸg.N\,י|AXk4`ϏE!mR]C7達S!PUNǏ** v༼wRϣV)g3+-}95h^ &9QI+u5wNJf2 MV9BGڱz,W NhQ9ǰx-a(깒ћ%vD"<4Y챖ۆggF N67 [ ۼ_U&f}_o4 +RV䌡 A[#E"GJNcJgAc,6AhQ9b0XJBQEL\8)x 8(>{(#nŵ*foMQK>Pyc ϯ5\Ё'jJ!FjNPRKvjڿaQ^^Z\3U49$"g6#̴ #= BhHJ\;f8:u4]2,ZBs0j<[N)"(R*q+?"D4֝!Z8B)L| n=v ]7kh!@2JѸt5UL|R#e'q;J<D6ʿe ?G.t.1#^hZ U [ܓ֋A]ߍ.a8Q/U8wTz9;6_r+LvV2o~;Lz'{;4|cкOz:EM0dWHsjhtzn6ٹ`[sr[c&j9wԨ/8rxyLfx8|`bS~]O;^v)劢X`K^l\)RA)-:ĈJ5%XR$;,%<7yISn?m|/@<ޙ)iHc'Ä7]Tސtnj?A$7%nX]i/9vf1|%gS4ag;خ2s,Mn y*y-z'߁v&3DE}1|U х9h܈InD?W$;J;Pn+\V)yiMU*Ts.SsG3\SM`!Hm+5jsOJ|Is^zЛV>:6g9쥨>|q7:dGUB S "`\0!Fc$T/pP"yvGo,O!a6'k Dnb~3^L"N<)G9v;ѹ9P{ojFގ,58F|Jx$yÊ0Y*T|L}uè \*ՔT'74f?_T4'%VTԚRl7'3[xS]K'Nh'@/ܔJŗ*h yz3:1Z`3rWp4'Uz0'{?URJI3Nn"MN5Y)a? ȓK "4cJ& h? +87,Y1H _y$@ZD.8Y,v& 7 |"G` " e*p΁_Hg7Ȃ7V;QpK z, 9) q+i`8FU'BY,`pD& 7c㈧Z!K"L[:ukx# !hLm)$@ +Q{4Fz5.4*'IPc촀Q\~t8pakfo'00; ;?Ec |L)t7&qfN0뵣cʹ2J`.RϝU'0FF-3?8sJvAp,%kϳuNxR?t,IS0c &}L`RY:bYDMY9 yncrt@.Zttxm'E(֬=Q5% ŧ>*|(SKjR"^"F]%jN]]]SO.= #>z/MxèTgձSaTH]%E{U"WDURw%+XDnr%jJWWJ:uUR0"uV5*Q[P+?wu#}*/$לpLR67Mj&=j:i[tVsWӉsvjQԴ\qU"RW\uUURN]DuŴr{O#GX^kc~ud ~ud L `g;C=y7XCB7H^t&A lԼtӴx̉]8 WHMϪzj(whFOGr (a$^iI!A-FV9/DKMPMw|bR,gWbfA9=6KHȍTx)I 6RGW7>ywu9qߐ5xTtTw?r;/G?^E˫a/wcct.'wZGР~ ~+y>\5ిx6A (MO:ҏ/?|0 yI}šy -aIIͷ0pϷӼct>Qۤk)9KaƯkr+̇ϗcw}f(9H62X5*C–t@Hm`i^ wE%D@ (xrb$gAD|X-Ylbϟ`*.Ƿv8:?Krpsc8v9 <("w`mz Lε |߷>wDrׅ\swlZ,k0hߵz*V<579E)Ƽ`cYӃwfvl\A?vz~mnM#dz%Fpx+ & t9qv0Mus<_b ؾ/_ii޿Zo0ob,@8󤷱r'93&w]W^NŢ8x,KZ.ZY4ꠡfڿD^ 5 Khe|;ijh6[ AZn{_8?ձS"1>OP fh/KGt`ZD"rR5 Nt3Za iU3&G[p4HXL> 'H kac}9@A ptk[chi@ QzHL z ",iP%LR#1s>jklӲ[iJV)WZW 2g{[`uZUpywY^Tq9N=i#I1U'/׮^BgRʥfwaG [?l&ysFԼ̈́P[큦;q+m#Y_e^X}yk,ֻY xЧň"$e[S=/iHǀdqgUuUW\۵I(;MG&)CNިKKP:S,zdC2M~Ϯ͢B(CejM7W^mlQB<߮靏yʍĞU*(N!0d\l"zuǥxP&_.sL9Fa=/ü: DL-ZBgT)BU%h+yY?ә:=V)(sVA-m=3{/hlO`̭Q/S -uJ6Jh%wed: i'F5aq+ܼy^sb)#Ra6f%*獋9Sԗf{kl(4K+R R3b t,xNP4A{) K3an(5p:FI*6^Uc1>Nab`&lq7I7F~/3BBbwԓ=5呪^owxk]ӫYۤT.H<8`*Q(7Ш&p0L8XpXXդU^A:RBRv X5FҎYFZVh8zȹ%]sqFN* \rmP&qxe$7;$7;6$7;$7%ҠM `0#AI䉱A"%>R rn# ,ȉ_v( YӚ:7G l=!2L=7+sϷD7N;:vDB~,^ǂ5ě6[L A1؜arϹ4x' SϤ;|o%v[Yh\??~Lୱ5`揇_ 6{0}}@>XoIU3輑77ɧ K$X`T9F$㑅H>HԔ1 A0<)c"ҭ \Yg6QOgU 4u'4^ڬ&.G>dG-%oadY7ߜ%`"{&CD{h"(Jnpb6$#Ŀ1*w.aU4=]oqw1UxmkAw{V Sxqewgv1sŲ8.NXS)ֆ:}]B/7 ꅪfe0n<޼9+MI#" ʮ̀xx 86{/=%Z(=q-Dk<x*r =h8sDqnf!wK´@]a4;3)ճ_k2y ,b.eEbd Jp (lJ ci ԁ*=}R@p"xNLjhZ$";|+C8/Ch}v5 aCg&8u\]4tb 7oՃ7?g8{| bzTY[4pݴȥ "Y? ^1ZX6-65Ca,ͼ̇4Iz'6{q*A[jX@V'O*0pR`fØ%~azR2BsR!7)QXkb^I^١AV& b>W2_̃? X'o@"z٨X<%_L,;WN+ l&v \tϷ;jηa>Yڇe v4O%pyU/r mU 1xx9Cs^h i5SمC9fq1PG:9WowTB4O<;(`SfqY=99 x$Xu2LwZ D佖IhhnD}?L7D}'!) ;jA@SNc0'z%E<Y^-QFkHxia9N`.0 |#JhR"Z6Z[f DžᒶqW8Zktkrkdt|_xEj7&"si*7f!+j*(LjnI}XJ9 %Z|TN9%bdF/cL῕'$ޥȆ8e br?qyɴ񶢑%/%(D%$@tY$h 4" bwre!\Hl 6t  D9 =riL@㱎Ǿ! #$$5pFBb%#11)ƝIj$N0鎵c Guq!9YsbCN|Q[8NJY:/-|]i-4":⫊#BTE Þ\8!sƑ.! N.8v;؁"vKᶔ %zd kKm2,2ϭ\((ahZ64 m=7TM--BIeekgY.-|ei8<Ȯ38~^Xnm~"SB+mX_e.c}_l0bKBR>o<$DCFiqg_jrKV9HmoiK;{Ws;!c8`1waʽ3ø)a@iNP6z.O;&3\UQ7)؇TQKQQwNEYѝh+[F V̫hb)#RQa6f%u*R# gYf{bclWHcH( 5#v(PP@byEԾQd w3lZq(jtT&8"m0RS c}J!Dab`9+:2^GuoR'w Ub p*F BF0<`QbaV\ lPA:NN! N piD m0k`€ jb~ K4`G@2\*밊ƥXEdJ-"\B6xјI&r!\q'2CDBh´Zmc+LR. - ::#K94Ѿj~Qړ|^)mkhw3s(CW!@3OKC J>eG4k"n*Az<>*}{;ڌ?\E>\z.,lw× j6"WqˍՄlz'q3*L Z2ϊq[v+=G*\(v4BZ_6B6½eZ]t5-Nv˳z232\ȆJޕ}¬|y8SWPT}71j4·d'dedU E{΢g0OIMmA.]<{Rs}<r,s 6Jc1F1sɭ`)$Ť26C8%ҠȅK8dž 2L"O I-*0nhs !M\IgFrd _!d|A4KdwmzPu"gű3vdqDïsDEG!k15<2cs?@6[:f l$T|[FrK p9XlY2INf}VB\u2(Wc)*e\N 4-NJ+=wUDDI}}U=WߔHpP@yTZԸq8i7rӭ{I@$}Y&#Fqm0Ms#Xpj ώY-0Er_R67MZNoQm{Y0,y3 Qiդ)*Οfc3v}:l?ޅIleAK X~~m '4t).Gg)14jw(Q[ <ûϮaA w?xLa}$rTHV9bzJ!D̔ט˯=,<NHð,\{/IJFBP39`aDf<01а:z~7L^`.{GDJIH:03,*ȁtH` O ~ ˏ%aJĐl925a AS$-wJiU(;-MLCO_k2y ,b.aEbd JH fؔGxH-3:Z-U (}R@p"@\^s&U4adA.WlΖ@!L~~|?3aWpZy]jrV]8_83 " >مIS)O3ogդHtrfu< B 5Q,kb|yMMՐjoY5$a0,`$euGgu/Yc^kȦZ۪:N>pZͤ9L G.llh}lX::Pnk)Kz{p&i~ǐ|f3_?A G%"\Uߤ[*`V6?(ogyL򏏿|8}u||}ӫӓ=9'Sqy# zc~݈Iܨ瓷ZU5WW xu:|z9i ^?M$B)-XOYM YuedЅ<ծàci Y)u\s M4w+7p^/|i7S P\!zy~cS3Ԝ(<3 nNKSv ~/X ;3mꖛn? uF /& Zx:7w[f:N<}[Egbv7qBDЬ G&xM"8֒C GZ("rìL1kV^ϟg@0't@ (`)w6 1+W9c (,:n+ kmN6=FغQl<FO3V,U]nl }aP^,m!)Dh6WHr\`D$4i[wĩ!h,x)C^1I,` r>E{Oso%^dh_.kx ٸH>&_y%H[)4&`6*1JJd:0m¼OǫjMj8"2_=iapw-v!oڴo 5tqߍM9L^6hP+R3Ey&zGqqCo UzާNS:hPh]Z[b*yQIdz&¢| oCv>LkϾA6ri+y-> \HAx#i˱me06h&%iw}SFĴm'pc.JDZ|5.ekn]Pn. M>oiZy"\y)/3[Ƣc\78Ŝ9s˭#:F ]ǥuݜXpӤCеL+kΛrSH^bZ1b"iFVFZZ7g#P4I&Pyr5 Em^[aTT==UHw]wZgY]&\ W݂iIΘ2Q`;'fN{)RseZ=%hi?B;=RFJX%yJFo"Р :dZrn֒{FܣM`R[1 trhz`kB8.04o+:rWVLN-*iRL$ _|5[ъi8*>JT\ym"tE:%|c3:_ރ_†o^nDFx~ͨs_a|QP&-pk}lp |V, LM cWlG* .v poVEm沥iO'̓]VZ\3U49FPsvuή=uv̴ #= BhHtX;f8:u4J ܁ 2,VMe9NUG EέR "(@IX  ȹT /HGJ-)5ʃZ_ˋZ~ku{S,XdヱEel-bR"~1u@*,DF"ZԾDN\=qE5#%\!D{qWOP\1$\Jʃ1zGJ<0׉'#8FBϔ9)S%r^9_]᧛I(e8!췏'ϲpRЛ08gpxPϱ gOnɖ ARabs..t,.F]kZeP@Y {uaD:+)Oky//[fP ^G*c޼<>^r# $g\Kr1,e2> Ks:;[r\JŒK?EKN0%: q`U"pZ!]\%*SW k9dy-2^fW?wME|j^f(cɉ$Ke>HsdYD(aZf0x4 w^I!9ǔF83_ S02d8BАR Et {O=A}GsN;Rh9 ,#*7(ɹwzO1 Ѫ2%i͘Ƈ*:0 ,`Mz3][o9+¾qNl``YJr2ί?ղZN jlvU+*($\E/5Q%6@H0du*E-J%9!U $9upў= ^le;РCIot>_t KZ b9T8 bT,Sp*NĘ©XcyT,Sp*N©X8 bT,屑"^J`R,%0K R,<- Y-y/yB[` l-tldyf< j^N~8*UP rHMyN T3]sΨsvX#K([e-Q)3h5vDXArq&(%]]MbdP xPob0jlV@ ҁRРs+p~0 qvoy߸v/ICAdҔmU~<"/ 珖M`>*@=T%c(Te`` $HRD:3U]< H([iV; w !tA[֕ֆs  R4;B)g(Ġ.UT^F|DER<IZig&v&xZ&uJXHc+@d=!*{TõN)cMJ&CcI3bukGU-Uf*:%^TqFb) #c0)͹ P& x)ExΫXtC:6~%Z~<mL5y[9 #TҁJJ0o)r+y\^3EXRKj1Rx?g:N})ЏsпYk< pof!f޸k 9I}M?n^?,W!4/9!qV/F;-ysi:NrE)ՙㄹsp MQMfp2V89/˺A{Y74KMNj?p,u FcS:ZE?1hXe pEFAd*G$*E2>~/H.DK=;q4G'S oph<jk\l׍m U>l=[!f^Mhyުhܢ׷;FCVfg{Z`d![[ /s\Mj#BkB.l :%nU Hf &{Yʋv]g>L!>, Av1gv<ХZTj'E^V΍~znwn fˊޜ-Mm\ۋl*U3Pryn5F9wF>$X1EuF s͝ ܀s8X)&u0DV]p3ѷUH(‰i㹳j#0 y!)e ^yQnZ>+\Zq'o]||a.v+< I\|Mr;UR⃡ H2@5 M%(c(< NqK:^a+doTil񊫊੒*s*)MIi/ᨖ<ʎDOV}]'xcvxȦ^HX:ɉx]畇soMc^V^Ho4 om&v~u}a?yq9MLf$1'00Wz8G݌Ȼ?yx>*vrCgYs0QSf?r0uw3x͝ծ?uqX1 k6zlf'?E *o߽! jLݻwOlwC`qy6ݻVy,Czf`f?fXzop휎z>e }Snw#W.QK,_ݖ[MqPߔ 5�diT3*z~`tR&- B4T`hb) Fqk!x@;3IXT ^;ǰXZ}u1q4&\DBt)p- %eE Md)[9lt E՚LZߞoEuv):}0GnD-m淸7H,'.Nst:/^3 [ӊ5\}zR- )osǜ1T>nxt9nXNY4Xg!J}틾dLw .àIО 7JLƞL%UPT6:*A߸RhS%BSʠ#Ui-n '3ރ uG߂%0'2 #ƘC26$.=}wΉ^9f'}; ׻{CZv$Z5!s&bbIb߼3&H e0[:w]0`UgVq{Ou\.qػV+@[5 reG>}m1 ,*F\zVΥ{8qS_l~́m)Þ[h/:2Ϯ5 ; eכBu'q<]y]uHxBcp%*:j u=Xgʼ9`-K/z&|Gc D"s&AS"'478ъa2X`jk Xm|\#f>Ο kmwP$;B=3aC[?63wKǃ!\10^P2o(:PĹHd8+e8xb4-<ת{Nt6((V J& [@ zH F _#ݕ'Dx.k.7ߘYeu$Y5B juJV>NI4X:)vbo-#^ #VZc<}>9)s]Q t0i%6:"+c%N(^2 J<585x%x<ݧu=fVO:ے3:4tq <Mn'Kis~4L'z67r&<ΎŶ9}۷b]VR?O;/=T[j7hs|#\pYOyW\>;A"*|3Q"H^;ǭ:''xO,zrN:;!'F!?\':* h+ 0IRRZjjl@ZXH$.MDe"X\籖L Y*g]>%}φwS7>`5d_?lbL|iz7It0LmV0u<(MALWWǫKlͨ5QݜPK9T3- XdEm<t8q5.5"Gc'cFY2x(BD"I$)" ZTWt MRH(9]gNϝ,1& ̌"$/+q#94t4R.xa^p@?JᢀYv! S 2 Ƴ `>7CuuZw1Z` [t.JNSmuJD3Vdh 6@uFlUӒro&<}ccK,src1\M I á;bc'c3"?Cګ5Nڸ:A]8H@%{J-o~˃[c1AEz3 /Y؆՟Wul/_rud'<$>[X8П|2yUv$s!xzQI&1Sz.z0ᕞU 8UⲎ$^ikpb y)]tVέ`kk78zgfJH`lCc>T"7l|Ȧ~^T'W=7ʯ\>l=j<n<vyk/W5$>z$}*'dֈs,"'ZDrpajO 8F|0p ˽*E;䬣'3S:n:uR/eR/eR/e5(G><72 kI9A ]b`ZDAs I AdC~:2m!Џnq0vs7SsV޸7dJɭZ(%jt,9[c ڍUBvcrbk|1r603M>8 ՔOU}ؿ/FXxM `Q^Gk 6aKA8A`Ql:VHPd|I3 o!Zꉏ܁(3:RXٮz\9+y/G౮UU/2O6𪝍 Q#fa3y?^~?YyE"%P2`@B6dr}UpiATs(1 b eiensZXtHP0=gV <ƠS/dm#U2g((¶uU[@|cVTܑAv8M/@v0}_]d?DA@1s2-K$NDnd1at l)rXg~$LTBt;X8Crt:z(SKJPB8`]H"`TT֖аgN:thTu",(NSV$\qI{+f:Kؚ N:M4MM;k kA) 93)+K9R:š3b+y6LY\υ)#Kk̩3e)2eP;'s>#b,2bF)9z):#sn[ +W}.*Kթ+V5T+4W\J+v6*Ĺ,nCW\ B|rVn\L÷TF e\o؟48,bq)%f?~}7ח|7Pz%o9܇qb~{k.^ԌjOD``ڙWh֊cFi5nbxz\G$˓>O9i~E󇫝/zsEW,$"1RN6HL9A|(Áhʀ4W!'!$O9G8ĭCPΰ<,T_j-zBʢ%:~ .F@R#K+$sQKATgVC_|oM}\c >5Ǡ|]9f:_v[ a7Ov){0?bۥ'TQ&U{ NQVS5.q?9Yƺ_:bΞ9;/{b"ND.)h ,*`LvuGmJR*e˔e B6~VPL HV/2g^POO\et XzI}C2D'R(U=& ۈ~:tJTcV@[k4LacGkғ9 b[_1ŰCV]qeOjj5l|܏v" UDV Ue\:!hmaa=X5b'bs=JilP6!jwQSDtGY\3` Ɯ67Тk BܠLGp7傩DO'@M$*`\"0Z1q$xd%$,Q8qO_B+x*gmrЏLj:/Zzw՘}F۹ax`jH>0%1RxV`MFLE7 WsXh+ c*D'|r270 v<"i;Jf9 wk<}%pYsr O|2x&If^rr!R"1DLyb,f _8qgE{m62߉L WޗS~vVMQ#4vigǫ)d ;VS3Ǭ /XھvBGJaN:'5k!;ӑ at?N*3AՇ| o'@"(Ym~~X L61T(c 7ui)55JQ0 Ul\Y]gVJ:qZϊd3~w;WIPLʣd 3Aq!crc6uNͮs";eUǛ sQF>$yrshH {cRJ+&Z*8A<83;<="ΔBr!fv0wEsʀ ŸOe?ʠnZU L+`ZABD#PD2(Q@.CE@p"?K+{>l'p C`uەo9j|/bsDJN%1G &ySc ĩo~tzOh>K +D2 _"E0Lw8ЧT4"Vcmi V ͢(~ΣB[ox~Ih4 .AA)&hN>@aBE=Z߻˛P?O~->p{a7Gݻoaof}퟽;hìp" *ihSؒ[oj57#dc3T^᨞?8y ߗ=]ٿvuNV겓M6*ju#i8}>\5Qh FJjk]WƿsVy:^äA}>_ dFEgM_д*4l !Mے^ JV nevN.,o,T?{u7+ 6W͹v-°6cf8fsQ_?AȮ{ w;ͮ 88:o„&{UeYcS7̚)shn> |8E|s6;\;g^FگYCsm7e䄙k28%O_ݖ[PЇao^rfW+z =-S㮑uAIhr,S,Zxjњ[c^wѹc 5G -=EDIRT(cɂ|4?BdFʒ"D"[9S=r؜8D|nGfCǴc;tSKHN.l2_$)8P_L%TR(a+;.ɅTɔ00Jh$V\i܇P6%3wQhH$HN,͉9!9S1S) Βԅ<&NF:rtHT [}kqGnF<+|S4Y*誑H (`G7! $NDm`*Q& p&qbYg=y"eGSƟ7,6gS^_} %mT5[M7ԛ_NA6!se8,v3^[ZXbw5rDZ4p@wwE3;rYwڠW7VN)ά;I6ްNs0]J,;y)'ԋҰ @Bm uzQyguVMVn#Ymz8(#fI߄wGtz9AX>ַwHXNJ {|Zt=]NebEo_XjE{ceX&Y|9~^ϝPimB@':}$ٽNvEM)oIv3 DJwe=n$GzIMy40OXzŀSj"$eI6o$n6ռ*uIԪ*#<ɚ]hL葃 e;Ji,Џ>ؒy~?\/z`G)>E\RX.o-Ǐe(pMM>z![vf1QsI*XG>w/s SAݼ014D`dXN(ןEASnBqA\Zd;C=CdB[Zjt^?(5\Uׯ, ( "w,`pH! &ۋZYnL|A+ACV͒9cv<^ȰFJ#?NZV ΔRw9 |IQKex'`+eg8tEȡαBZx/^RN;U \Us]gݜ?m)\U+UGJPvʿ,/ /ܤ Xi H? sJ*Zdfq27ySsZt`$%JȵɆ{eFZB^)\b}~QX zB:OKeߘ1Ѯ?oc9zW?t*/IgtV`ԫ`2UkZ^G/RU5ECpUugd%^%\9d)uFvZJyMrP>i[*9\wc1xZiVnlj\Jpun VX \Us \UkѶm.Ջ@:Wd63pU,zSi*h;\Q|zz=p+2XcwચkEW5B+PBlL6y?fcv$n> qyCJȃPM/h0/)/ ~*y<me{mhe#oK۝B^5Bl ͇ ;N.m^ێ` >ȾȾ'0-wIV 0ZaH j40|j3c"wfi0&Fk5@DV%H|PTx͛5^6# q?*aG⭝cғ\/ꈾ﷕0gOEmS1FT[q䩫5qɓN\/DtI$0Y_Kc4Ɓsg8j(tc[7_C2cfu{ʖᾜy.jCujɈ !yZS/ˆ( 6C<$(ͧ(I5(KE R7S uiAf0 T_k8bcOMl2vaFŇg./\9 !3/ ,mW =yߧ<|L/Y|`5%yoKy^ѡمDJ.);\ c vSUClM9/|&l&F@(J"- cĨ'*=1s{<x'L$KNyA u1s\Xbs8Xo.W)ɷrU֭[?[Ol?-"?$x5xj>|ΰSiJW6riq:6u}G;{KfПX*)*=ElSuf^~2ksBzѓl[͏~jSoxr ߷Os/^M?.H^yOF85_?zx)v-=Þ>&c_byhՠ khsf3XS|AW'\ I;$" :Y$&oCz~-ϒE"]c1sp6;KznDR.aQ&[ce%ammY$0jE0yx"8<'J"1s.IdM6TnOVSՊ';Izen/MS'T (>%0+JkUy 4a&iU9xnp.ʺ~1p{Zri 4]>n3_]o=1*fcs"jebs@HJUg$ei|\#tPI'CThi7ծ$@"oI$*Y,bÁt( Jc==="%hSpf@ca:S ˽Ɛz u9jjuM"=.5b&kMܣ[Q VRԜ9ItvYy_-:}c5&P"\r:'&Y!C*]F-K%9hrc-ii[}UwD]lN| Ao4 _1C mXO$E~^ `?.<. dWV.M$f@"Ó!nR 2e-\b.dQ7:+p76-@j}Zu ӞM-W lZ1c1OpLWħ˧+$gG7bG( $xv'7YkEQV+%q_a jQl> t.$XY2P<1gXVs<Ky#g @Aln}!YmKd2XCiok{ɼvC|ȠC^M"یRHeQa!eQl5RPDuMJܢA$ɱF?2ړ|NZh@@idl̜#Ne)7zƹXBcUI:7&-o)wmW;/ˆ(Y kX8Gt*䔐N'qgC!JVh"UIrT¦&j.Hp.1mz)s#v҆l j7;zlk *K| V+vʆ` eFMĆ. &If}sTM6wzJǶbZ6@B(#(6B%TG )D Yg-6FMF@!B>czOuî f_O۳oVf%-fu篎fߦO!zYiB`%%Z1-b[[ OuBYP2ଁK !$p<]oGW}[ !@8l\qЏjLjIʉg!))ⱡL񫮮N)g -e_xb֥4sE>MN[^˭ *L $1ed%Q$J)JXBP"3"{⮴G;f>GiLkAhZ8'`1T8 RP:4@!#RG> ʏM 4 d3>rn "j_L`Hɉd0$R$U0pc $y8u_ϲXiyOh+D2 _"E0LNk,p% R4=#fCmi Ir.w@(G:V m& 4-BLʽ7X8`B>a"{7 ~p7&._;o.^+}x);hp.gy|6`4x3-tE!-#q[G> ڇuVü h\=T8_ d/P FcdMF/_*'?;;̈́f ~ڭ T͇P˄gLˢǶ0kҡh̼"E[)W %x-<~eiM9af?z\@h%O.ꭎgH(}+wUlEE;SN##z GF֎HB#`bѲS`V H ňqANx#v1,TVUs.:~<3( D{&Qh ^BԄ%E;9Evr0J {5q` t@Cw(CzefU<:L3TK9jq_@$)8P_L%8rN(a+C}͡g9D$wR%S8Kb(rXs%1rBJT CBC"FFrDfi^3%sx=b,9A](^9$qB]54x%79=?jݼ΍@:zOwlR`F1Az,eǧцM&FJњM aOꙷέ.#<^P-gכo޲boǾ3a(_TW{F^,H d$UI8_ Bej)a*qn"NAӅj]5oM ~yJ>e|i` _^fӿ$?_T&Q)Mȧd85ċuS2N0p5@CK^&fS`n^)4roҮM_I\t"ìɲoz${dL&topI?jۻkZ(»|||Y=gÙ=c|{mgf))Zy^˷S˒W{P;Y"A*< Ćy@d!m[Be7ިf뒔p>m%[7[ǑERs,n`xZsŽ:Ykwl6v=b\O,vΜ7<2I<#-4o{y` wcJ64Yzr P.vʚJk:Q:魯C&tY5wǼ'|__vn,U$K:$rHR(R J>H8{^!!bkTp *rx~A6Bz݇?S=r2P"OEԙ5N~0c&vZI?79? qӷ7 "'>'Ed4FF{s}3ǑvG&ǥv;8*-[EM")ˣ P;VʷYB8w|r|5d u 3^Ȧ@)Ւǡ󦷣6o3ij$~* sizOͥ)1h:?.`wMf26 y.& g`͌LKw\Oɒ|fw}Wr偣e>X6:E|rW(aR,p>(jDbJ8+1S;O@?Y)QEI +p\p ^Elƥ -MZD+}JRRknD:iAJ@=ƘVTc#g7ܔB]:;}`RGs[ʃI^(ݯ{LHMknjQ#RȍDr+rIhdr}cf>cɞff4̮4t<$0&'匧їMr/i@T$Rͭ %U4T O!?3βJ%Q3urM J ڔR)Enǣ."^j싐:i [B9s^Ͷ9sڋ#0ŎZDXNiqر=@g $Hy@&]c9F!*$2Ć*ZB eM}X1֍zؑbQhY8f6_"ihCjH"JoTg"ɔܖݩu:W;k.gJZ1SW]lNLF2+lmODo[WMK:wjnMWKd[rz[[ݝ mX{mOby;:>Gp3ص4h۩ϑoz9z}9 =R/ȭσ5eR4!rs҅ o` `U_udfxMQH+ :GKQ9.$hIi !yoGM$O2wxi~(rNS_7_X"Cȗ_ˏhtZyp3LpX/"F[TS#9|x 9|x>B%%j@kBk𿕎hU:M1 Fk %E>He,謏 } F91V2QTS} MrKΞcg byn{Roo>1& ږ^!$&Cgyz6`jsˈ3aS37j!9ag#1$e2HQ%{[O.9n=Q7BjHD&8D1T1)S(5ɠy^ |dn0"ܒH8F kM5z$Y|,16HEX u!:ᓓ i^D&Dw5dLwA|g 9:s>@#8@ t4xkU'Èg28*+ I9@49uQ9m^X"%)C$ɔ'F("`9{㽺as'υ2%ÀO%L7жfClٸʷ]u26.&,P@U¥#'KuhD@hTy(~&!2F"ߚ 39"q-Mђ( J1r$kNaL(}RQj SC$\0>(GQHqPP-Dh,XF8$(Ր0kvH 3:9,蒌T;oNU0)ŸǺc2(SHD)˄*Qy "$HI"gȴZ3j$J -uj8CKSw"j9ٻ6dWvcuAE J)R!)O̐") )Cq$g8]UUu] 5%;Qthr, ii1CGah0Ű ϗKιh>uLPCF8%8ZSOr+ H$oirYzpOz/q@;hY)%>0Is5aq pT܇>FwHٍތ2੒*sA gO&"!nE-0(e9@qI;${ʏul4VeZc@v7g & S`X9X/6,DH3B89Xee/4Y}`9It4#"),s<ׄb,uU$( #uHiLd%'w 9s8&0U[8R!gF?'6}oi0^.r{iv} ʻ›EUޠ:?dpi?i4< 3 ]Y5^y{{aNj[8~1n:wt6 ma}BO3g1Kxe1ѳ1uTxɶQ[H|3,)Ys&}1hQ4^Lj?7T[2wXt8|?M..A]$fǝAX|es4^ac76UmyU ;_+^]?~?߿9ޝߞ_`]g}+ fo~݋eWCxAgh[7~q[?C͊l$B-7QA HñXK{un8Ոq\Ǫv0VFYQ9FޞV0.}z?sgٴ· LmWDqsV z{}JU ]od>:4ycFYPm PlOnVuK]CCC[[G9lx-#c dq4T`hb) FqkqĈR̨-IXTŰPY➯}t3M1OW[2̊,+|S|s0Chn/ y?e~ZZp4*jxQ˧ μ3 c<cnXօC[ȌU≙bk7^^WL-[ګ U.RC@1Hv!2O77~\@d~^\܆ Hiq$eTgIjA0Ӕ_`^F}qoQ'h2mَgs7]lcbwCE @5wF#;1o2UqdlhGGZ8Γ AwD\ KP$uwgŮ(,K%9Dhmd3r0/~2 yLlFA«k29#;;{`*r?`ͧ_#X/ pgӸ>{QwϽd+΁/*$f@YePYHYدSKNk_ >AV"Em%]ܺn˵[{M֓.pƭ 값 yGշ_WnA/3o1bW3g=oi}w]Fjw2tIj.>}YWcsGXN*[r3%22 E hd;5u88z첄I4B'Mu $d1 wj*3!Hzfl 8 TR&ē\x 9Z,] uvDg6AB5 B]>ܒ=[.}RkguY姰a[GGdt'$0PPJCI`( %  -C%PJCI`( %$0@g0eJr[t˙nI`( %$0PJCI`(-K$6%"Xz,c.%"D1I}Y%qaχK6ET_Q4A8G;%rvdy3d=™{sm0&\G?#,V$0a,*jD%he  V; Dx228X\籖D YgvFn>K8t9ZYﹶzyD]ٶO U!U;h.@LG«LK}Ѝ֨yI./%cad,v]"?a!AVyX Q(Z-cY(魲(1R3#`cQ'ElmKbG:N8`SIolV5SQk,H*\ ư ?[ݸӏgd4_Vl(sʔ =i<^SߛˋgnQ!ֻ/-}81ԓ*#Fa`Q FH1IHsv+X:τB)2Jq>Ui㹳%D!RaB#I;ȹg$mR8?B)g/8HE ]Dc+v l0څ$ D-6&M|eh!B"cc@B.'DEG*)0Z%k1$QEhe DŽX*8#1a) #c0)͹ P& 5RTG*zc'㴯Oh^~kC[T܁Gr D.6?@B;sGBSEqrQ[$\P{ s\>f|{Y'*7Q)$-W8P*GCb q`lzVMvC >5qO~JOmu`[+|lw['鲻B;+eT Il89BƜBr9ͧW.@ Ju(Hg8 G^k PnEI/U8\M=o^,v֛gHi̊y(~>=2D9J[_`g^HOZ$0xg0Ob&w$?JXjG4w_zO꿐 #zJ<ñ\^eSE@޽Szc>lC_<}X C/E:a%&8a.x!GBDX &g F8b|:uݹA-9?ɍ󷏭<^Sðm,8Xh^F1liٻ޶ndWm\53pHݴ{[b. q$M~$Kvt$Y>D)!r^ѥl1{- E`}KE(OT{\J E+vN,hr\k(rvhysՙrǓy_>4w`nz|> SwPs}$(*߫w+3MΠ-9hؖvPdVl!UcrըzJj}lnP:  U@jhd=⻷j)ςqơXB}Gܮ߷I%N58K ͋?o0 /Oc[W*EFܒy|5^GIr!*HJ YQi`SML)[峭ө"_sMBn觓ٍq:k屠`qBEaV}txrY7Uـ#qdX5FReP<E%"E3l%bkrĪ(Q'(E Ib#g7vF}Q `q("H'D<3-\tE]y#_ܙ7龩p7wjs\j]bw66{2Y`Vqt|5Y\q\1$)La4ja$yI.>:;kN%zB-VU5UUE,UŘCfUEr٣/3t'OSa-3}g[nˡsb*jE)L+ XPڋd5'S) Pʻx.S:V'UsNrQ $NNN!N(JhIDP6M,%GA}q^誼)2-u*YXa/1I0F_&9u.ANŒ@c9Aj xbga+e\bbǴÛadcVr$tH '-X]1'AXqݲhGB蓖[eB' @ސ鶇,()un[n׫c](}lR~?*~/RvՇ]!cO߿>_݅߉K*]c<7m|g]WalrvQ&&4ޑl]02 \2v#+Vtrqг՘wm3;.'7j\=jY\-;FӫԺlZ- Ϧ׍v*8w׭ֿ'l^nӵ27',/ə?76Rrj~Ҟ?3HlԪ`NL!-MT.}BGA 6> 0Ujծ)M9R1KS3X'rԒӽ 6#iږ~&<6a؝i1Ӽd|L|29 [Nuph,*%9JTtI|'xW=sB ycB)%*S`rTHRcLx*dRB.4u=o=~O 8^<3r|R/\}nB>XaܫZډ~V$RtZM2e-_ R:`0~rsL ^xtyu[2I ˸Оګ U=n:6y?]`Yz8G0߹vϵٌ?ϼ4_g$򷕹Mc\|} \u)h1v\-B9zb `ԌMeMs]ޅb7TqqsRg?̂P;/ p;`CS^̿[|I;OrcI. Sq5Z,DRֵg^-ʈW@wf-]W1DBW4d-!5 E1ĘUb%ji!:K;t}FiAؼgoCPxHjm)P%jgtKɲK:$ᑨido+14Wo"ۨZYZt_FNwC>^՘lD$vihtlŒ7)ȤvBF%!amliPYtj{YŌ9FGHk!l$h!bV@!N[B3AaW%$)PHkYn#גuwuwVX5nD!("n;R 8'@&b@fCv4l\FƑ"Y! L dt ¿Jt"T]zDzgS d6XdcngvEm%˲ҡj@o]\Q?A2n@Ơ) uڀb(G$]tD#2;clQȌ:B*Hq:,,:L;"p`L ӌ/dd@р2Й5TXVrlL:v_#*A65T4gBQ!]`(;;_ 7C[KQ@;hՆ~j2Wڮؕ(ދ0x )V7gT0[QзzEGkIk ,D8L=G Per} $C aUhޣƻ ~,B\FI@<38onp}3"ҌY7Icc^͋BIJ!䄈mBU}g'C>!+J$W{H Tb0P&SAyCs ~>XA`A3s՜P$rDMW2MʴC.k<<@= %$dGeYLZ[]Qx nGep6GC,TGWZD_Ÿ V :@ߩ` vonҗѸ d}E 2FM&ڗZEP(c G]Jr s@ Q/сwK }/$ճR"T2 {(%V nK)KcE@AK^>^B ;!mBxctGgcy:(%١ c~>XT34Id&J+12!~R*qwq`DYUitUrC`* a!do(D?Ft2%VmZbAn4C:eќ0Mh%J"-A*5.--Uѐr h P Q"$4y6*a*0dkJ`=ރvn相͇.˓Nʥ^jm793IVg}"[&n[fOB=LB` DeoOm5Ek)DzI Id5I; OZrm؈-ŒؓU %*`rh[S|'Ѝڪɶ롃Uf1uAAvPXfRhTU")RClւzD-ecW,Bv=+Bq&SK \sD`|?t7^^V`\8iB)E#"Hmr3X\:2Xd 4@GJFdU{19Mk.YyF+kh*M3| c:ɂ9 w4';!Zd'D +pc]O39f|j? 1'2 =O&E7&@ec^rʶ}D;ҳy#>;[|N='?9קӺ*)[i׾}jg.qIe!׶NoWYP;揳gog3Jl\7vY5= g>ي2?e7BuG~o[NShز4^I%Xs͚k\5׬f5kYs͚k\5׬f5kYs͚k\5׬f5kYs͚k\5׬f5kYs͚k\5׬f5kYs͚k\5׬u5;[Gl8mFMmS٘^8@h;?hԤ)"N~q:2o?q:z-G:cjxC+Ruꪫ1 YX10{֠HT:|^Y8F F*3{֤{"ƃ]t3aӽE9ylwU;TxVQR6xu?CA^ u~E{l4ÙT>,x󚐼ΦӏzWejR戀zjx)?/ݐ,evڟO={ue=ŗ*gw(d8@XEH!R .eVK/E*3Cmd"sO7M{Q"B }ޡCxDOo}N/n9FoT^A4>_y^=˶ʹw/N5W5oǻunrmFΑ7&yɵkѶj\؄:NEK:G(alO⬟`_,}tϿh׺z~Z,/&V>Q.ʽ%n}սY s=)-ӞU߻;!N+x"Bp^ۉۉۉۉۉۉۉۉۉۉۉۉۉۉۉۉۉۉۉۉۉۉۉvbHDA@5=&e_c"zpz\%{T┖5_LU|$2-ˋ&Ua٧IdJfG}$G@}$4 5 oEF-84FXhs/y2$4m(iE2w !UٮBGeI3"G-DB!DG^1>,6Zbs؜,6g9Ylbs؜,6g9Ylbs؜,6g9Ylbs؜,6g9Ylbs؜,6g9Ylbs؜,6g/+6#S^؜ XR$CSҲKI _뿦Tם6g9ņs{SWR=EdATYJsuyb2[^}ND͞E#ʿmF]=*[h֚MvڸMMJ߂&J?43H>03 g{f܌R$B=8j΅ o"gJDi!S=|e<]7Y7I[wI˪5" vgDmʂAijsQ*wdiA3KU ҊҦvTb=0UW&6]hQlyqVCzgmFGsLs6P\dJƩyk>t kS`IkiT;O*Ȁ!ҲC+yDq)\%Igͧ48 g{>\ş6J}3 gDΈ{@E=i y`\t G6gA7H #XJqX83B\49h9/r^<蔵8W7z *n_:Ճv^/wFod-iᒍ蛍蛍蛍蛍蛍蛍蛍蛍蛍蛍蛍蛍蛍蛍蛍蛍蛍蛍蛍蛍蛍蛍1~oZտg}W.zϊ&ɗN%G^/9f{J8)ؗW%3^ǗC? ȭUVc&WvD,SBVbgr6 9W߫^z{XhSU$2ZPBeML7昭 ieFDZ<7>PS1!c?VU'[_Uo=N6ȍ-sD4s鮟=҉B?y{a^=?]=|{4rMopz7ꆸr•f6vS>o~EVv+7b?8;\P ϫoqMgͿ7̃im]/|[mqt9JUMXBx^#+Ʀ"otġv"|WUTҤ¹Hz̦GPs&q, <ݒ+I+t}$}/$]IAң* QDT҄6 /# ˞-x\PyXgF:\ g;IˇD r~ <}{`WX9NG-0`z=x֨xuF_aNw^})t> dSL"O&{ ?l'졁<.o`Kޣ2l'e9]v]nI~.`,ݘAcl1<%BTQMRFV/EQdQ"CuedfDO30FU/S(Z2F^ZhLn7viepWF0OjGL`}W V8SX@{k<ԝU PutCŞڋ;M4%8 "iA$D,bzL.RyK"fʧRr]p{Sm3>M7iYEQ!1࢑&ᨵF!U yN)&#5 $/m=-##ck< U. rRyq );s,"68Dޡhy/^P{E8 G*r $ E`ڭeѳ:܌T6*Q R˂Դ͹r:PŀRF :To9,UJ0˂J8_2ٯ7Yna>m.Twޗ7܌/ ?y Bgή^qjS9g3ofŠf\na^bHkPCKƖ_US3ḃ0|(%QLZvPMz4kw\nUV:dS}Z%oc8 .}ll;Ưp6ܑK?JfXB wYJ?g|}SQubtpuK?]rNw۳*YHH;SO^`ë*{o%Uܧ].vz $뇳wo߿Kxs/ߟy uv?g޿=qp)^5 ؊kѴhiho4(um5.7{W4szծe(As:zaV)^O~sT#w䤄 U哓T#H瀃+u<2bS7Ԝ*0n*S ŠtJ댻s<f Mmmnwޥ"f|KR@Gi0)|xwiߩ!>)r{Y~d&w63J@7xo,mfmYAWpcmp3 ^>q{Ǽuw ,sE3 {-_˕Gfl0q#7VSyR :ˎj= "OD~fMճ/4{,v9$XǍti`C epTqD=#X`TtG"ʖa^j@"bZ0+Cʀ=[6[#g͍5.rQڧBx3~"]m0W&ϭ\vL <DPIn䣁 \k*oԏ(M#d!̩`{bYI<8fS㭩;{+c] fñVƾnIV|rgw}4Az$ZAQ`拕V+F|stޗZԖ^n{\hO=Ӏ>bUlF'o+vID:|)?Pk^xf:' 6XIzl0(hH + Jpb5@%y(3)nB[=פ?޾fKν6(^D5cb55Z 4XL: G #` fR0T7 ْKyW kZ ʷFR[ą{8NxΡ??}㧠1bt&DfrD%Ux): u+&F]%r?utP1 TWS5 ylY9ןdshD* Մ' ӟ`\=a4>cdz ӯS ЙeGQNP%+ RAr2*"4p=s^4czι͹`1X>^66:-J~|7`y|E;Mr<"W&k:2eWf؏z؎VlDK##ea15^bK7zd&b( h!=֒v wltzw)#Uj|xa4B ߻r>77::)?ӥ?B eA`yʌ0Z;KluL;۪;y*< c&T hm6*GT KI/@[2-Kpk<,Kq `|m)JCs]YdY*A:nlF:AQCm,2=h_X R;ĬWGjC0Y=UOc.*5X@ n -څ~wvާd -mq: 4! > R>bQZ8I\3U49L^"T著`9IBsF)=Jb RwBתrSQmGs+), Ƹ#@*JJCR-&9{@C '֊}O}AgymA;< *}[C#h 6f ?w ΐop3mDKsll` 6R Z3bUs/.hBJ,lTY@MlFQc{Qm0RS-yA)āH",RLpSz~Opۻm0^t sޕ$鿢yvDi`_gbFaԐ"TˍeUq<7^˕+$3Zhӱu݁wWCA!YG)[UqDXQب!g!FB `h2zTW$m*yJB!x֔-\p2 acj+͖g ] 2jp4Eed'G,QѰQG6Z(ё4K LL JZTɝ&O~MBn :*a-N :-E +0R'-!/(3ީ@ZVulH:ccT1^(7})eCsqړpQQҨcұ9 g~׺C@owmb[ 4_BwؾttEiJx^Z٠I\. ]4J+JaS=q |?qbFa43J`,B$ 6(rق#!4 ,Yѧ0>>m)\[ vPۜ>̭] X9;7Ǒ;2Fʫ`+T{˕T{5ZWjo?b7g66yMOWѦ>rD|60}IZ ndvfeݒ1 M+Kڰ YNOΫ-pυ:&tPɛN>DbCoEGe)n)n)t93= ECA*%ѧ kƒ%ŠU&8!5ICBCr?2~g-mnyv ]}[ ~boOJ|!?H}^xҴ/^6Oe?ρŰS\ өZuX?ԡR^3y9ߜ>otDOBs 1zTPۂd %/6n}XA;̥&_I8^Obd:99NNx N% *9 +;o%Aѥl1{-R`}-cѰRH'gLA$L"eT:ǶW9Z .j*J=gt;N)?{j荊yYwB!Nc ^m=ڻ73z} Wd:^s{w;FݳZ_EwX5Xދwɑ9 _Lei4(g6EFS #Ԏkƹ\$ J.9ms]O])* Z -g ,Ed< ec,# /WT\ܿ}9Tн//dq#,X4  #x\D֗/˨1RL&Grbao{xdQ¦46VlK9Uʓ$*k}xj9;Φq9n&#j Ua}tN{咉S4lY)Xʐu5Ckg[AJTaaRdJBffX$YƚL kdΐب&Zwix5K|>Tq(L>1"GD*0 HXJ*Gi,sdhT6FX#UvC` SlV}1eJ312h%ؒʲrJzcDl݈UGER!:Icq4E=∋[{ &f/8VDdaL"%@ ^WiEF\<.IP!˔Z|]iGcaCC$wGō/ɠxpoq5RIvJCeP8Jq?f#e k(2C'7f R GmR$6 `/0#Ű~96c.X".m1K/FڝJ]G?2T*ۜWw|unyqN;fq2۟f? _kn{wFxJ@{|0Ͷv{۫{G w ,s`حRR4׷twusez>^d-6!\m2y8alB'ܙ-j[E{6-uJtQp4dˁ''r@FEJWT`1K-0 %g3Nxv>50feyS0 s1t 1^ ldi}|j9{r^ GHߵi^z;YV/Y&a4lijzM$0Wx$f/W^Q]i^ԕԕ~u*Ju6؃AbV(u3OXh=@ˀD.+ɨ3de`!6(Qږ0g$mjXUET cP *-AmD(')JPRr͖۰dL7n$7.;> j$ L:ZQh$+.g쬅lE#b5fT&ʻ26}VwӟSn~UQ~;c^㤿n6oXگ\jJx(35,(]Q!v ENG/87́ag]8]Us?mdTR:`,(&l2KG"AFoTqE4I: XٶXP"l.)ܳ(*]K%LA($`7f͖g`<Ա~S"q2貾g&oYwܿ~說% ۗ3{|fQaTfcR -r>!$)bt.i9"AJG`S*!0YD-3A|F+,JhHkHG߼ٜ P_3tG"cz+vl^QvXM(R)1 iAZ#[-"O!TxAw #ǫzVsYeM0P1H2Q4TzDLaZX~ޤtv_т8T2kQ"E-Iܽuو/_V|0E)V[阼%v5(ȸZuU%CFW%0݂/衕l<4nOj U-C&vZs=Ooϵֈ=4sݲS0d| &E%:4؜ul'0vz[1QbnT1;ӏ ~KH%[EV2(I9cV"`a ]T.{m$gPRY~pJF瓊kdx$l=!Al9{\r0'sלu%O~1M5_W)T>_ի^2)\b+$XSB=Pm] MQҤJ,c,EλJdV̓_uK' vʨ|6[*MN2 $ePU(1: Z )Vݦ P})- 3RK+^Rz@_LV=576MtusNP.":M?ִWE:&^Pq_{dUr(sJd˶HXe=Ix%'c"â\f ,PZrsEW\ue8,dє?Gnj4kB_/]>x}Bȟ]/<: NdΏqt6*0]_rtBӿ* 3IgRN7M#6Oi\Mez'XbEA/朜NηݜVg}MnuzW.;>:fllQҽh;z=K_/|7,R=\~i{-:4?9g4.NGi5DJ-jXRwy49;f%˿ۇ헟�~ /CO>i)LmZMf gjW|yhy潭^k8B%Њ?zOgUGnn79~F0&F_N{0[VeY]-Ԋ] o.PoOwQM+x!Z#n w ͇%,Dä_t1糣ʴֹSnF ] cb|wOz1N0N9iUl)}.YF8R߅eC20KˆE%KrN5='`Vzה1Eh=aeu|a!Fx)G25+dw-͍0 Fx&:§×=8³ݘ-F)TV?K([5iH̪J|H|I!P\6UTV(& ;9MvrzZ'^Z+"襻Q}:O.&lUC$hi6ėֳ߫ ޹%e-;zzQ Q ZX/}hU5بj=BKpU8+A8#;s6pv{vU4+I\U>곁jU> \)"aW\e`kW4j)֮66Jm=>/rnlib[>gpD''-?+?ۿfW7˦gcd[ͯUY_7=5jjZ=V:7iPsegW\{6ZpVWb]e!f%Gj+7g~kG蛟sx|w_^aÚ?w׫ѿ_i5Zv|6aX|Z,SOjީ/QzP/uNSXrMi>:+U::@w!k$-B>pﯣ/W1W_Y{%EUR& S B>d 6% $%028MS1x/#ȁˀ's%ml+9G7~H=bX+g|=gumx>Ojsqzuqw<ɾ5 ۯ I~p:49ؑV4| RagQVIrjFE$YH$>91@^CNAy9z*v2t1wO$B}zzO k[\el칒ۣN/s>}_>OG it{nW[oOW:v=vt}aM=KB K b~7-_$;Ж$E4.Cr`uPSuKmLީtTEk`ZԐ_kEptM睟0Y&S} pϫL'd<. :(ߚ{[/7[ZRJ^*s 87JR5J4!!ISӢ v^ _ݿ]~_{ug }"H;msײcsTTi{J=JH^$6)19b tA )*QHق$NS=f)JzABOjk]# 'ǤWq=Ϥ)&'d19,MkR`]=(.f kEĘUAb9DL@uZS+sy`LK?xu}Mf~BMMVB|󓵦76㎢ OH}f.ʟ=֫ڵws qj V;mM(auJ!X "FI"+Ѕ mSd4dKJ% yT$ )R$̪ A526302*Ͱg ec, -v2&-߬uHy\nLn'Oز`DBa`<tj #JI:n(FDh\EmOdB hT0*Y-~P.+gJֻ9j6B̾v3P{`gy[sH;E1ެ^& G,xe#Ck2Y/%*aac`J8BffXWs2֤ ш,FfOVR "zsA}<\/?^1"D0 s9l(C!!hAAlG8k%cQ&D8R#GBYufF̏:0.λZg3/y).6(.W>OBdgL""%@ NC"iR.ͼ/xœ@}Bwu1G\ ꫹ҞaVkV+7G!0/7{v5Ybҧ_uc}v "u U(^cw}ʘlcR}IEqD+}fc^&GoGFI%*K g Gi8Ud|-rBI9f# @h8i09Gw|W 8mlg0-*]TKƑ4- V6뷏{[]u8fK9qD"b橪PY@ڃAD5f: X, &(ԲԔ <)u6) ,!Y%9H󥛙sdǏd))B(#c .R=>MAIh2?$FkaBmP2J:0Rlox(Э!tC]=4M൥6WaSD_ [/NNh}pꗕ/W w8>1  r.9 U[D(B*]R6Cyc=cY2?SMz6.U+, {>Y Gbٟ,梕:C^vA a >T ᠥǜ] / (1&W2YԒ#[”dZU))˜0aP pX$GJKa@@P+u1f&Vʠ~;߻БWߧIz?> Co~2_o KuơЀR21Y >F"%䬭1m2xWOcWkTqf0dt ?Tyk,}"pGi?mq:)ߥwZާo m1sv 7?nXWvjmg'6i8>ֵK~}g&7xW=ܿMVjLݑ.;mF_|)L]^xE 2{H)C瀽Y#C&EpLlfd/Uy5Ϸhqӟ8O LjL"9%yC1ʧi`޵q$2?I+{cb_`/>ɐw!9CvWw=WU:2!ْKE> TmK6֭ Z ֿ".y[`n9Jb=\>H3Z;d8y,OG~sXyt)O |dbǓՊ+w ?$]p!LE9R*I[(C1ŬL# >W!C'35^Hz 8'ate6M ,:J3c/@v^pZ!)#5 @@%!U)2V'.۷iļ#R"%ǤU(fE9)_"̓H&V;A}9"AJ5q )>nnjp,"#E#<]-x}"̥T<1ۭeѳ!RZQ~PipK@JOJrv+΂YdϩלImM"[eାpѭ.=pvmtsq9Jc?dku_$%^]*t2D\{^ Y b0| Ll{uYWM]xF8!"k8Ѡ ^HA)LڽC`FZ9eDYy:0^B{a?bH y*zg"cR)g0²#>䴲SGZ5w8Zn&9G'Uffaη9#ֹ@60YOn}M$u1[8iIΘ2Q^sԌpH81%CPK6; 0xP0V/%E=W2z" <(4cOq.<$#5 v\KoWf!y34޵EhJ^ǿ>:8+RV䌡 Aܧ#֎$Sbr1l5Ѫt!hܯt/"+ F刊`) EACmZ@(. N~=겥.},) r_q,g)@+[; SqIrG$"/^6P+R30޹6>閠rkFi2K.2W&ZK{Ok׶شKU cJ#HSZ;^sN[{o&z2d\ې]ϗུd~"}e߆I. |ҿ%ow?v{ by~yh]ɂB]03x0~s~Pyu5hMJX[hg*(]lH!-[CZU,us[Ƣc\hu3tny$XLjPYB]wkW;CexI\(wޤJ U <&՞:#,RF){JXuW5J|.yO LiC+ tCBV+ |&\&ޓ-'OҐpT|zTNI<=z4÷7"~{, ~+ lF8zn1وJVEA@(K*`zVS3Lӿ stZ"'6 T$2eaIPY4rmYNdZA<.X,mt\ٛO5+?ZqcFfEnai|OͫM$ K2xI& 1ڻLHz:2ek##scB틹J* # f`"H0\%\%r%sbP19Gh(=2W`\%r8@w\%*bٛ^ҎA' ~.%I yU Vy1+D!!w]?GLj#Mf"W:s~fX :^ 8#=}5m,v2[l# lN;Of+ &3lT_QvU4'Hi%[5idҶUg8J9jjYReõT} '%,d_Pu,d1b1F0l9#XbW!n*5NuW^oxg=AWN%t♸MxřA]6+24/(p-ekx_d#.i@\9;kAE>ג)E 0}Rnsf1Ѡk k [F2In.?qEO߱|qMMd)B Y:aT8צAiֳIͲC>:YCtݾx5]KPpynK譕q~Qד7DlIזy\srFi ye-3დ%']'NwI9UaVI0J(k@{S(sYZDl ha[0`b4'(xAcNzkv+K:]A]6Yx9XW,- =4L|LV0u W^ՂԲWomU' >SX`sD͙!>yl4&tB-ocUMp* XEL!,ј&F$2Ag1qlCRzb=PB, WwCi GcBMAg^TDɨhmr/(&;ktFi:O[Ύ H4'_ss9[dW>(]b;í޴7Fw-;z"vĨd耤Kǟ璩$A!3iQ ^eXތ-x8c0ny(|gxQ93͈L,Rr]wD?.-BCTtaػj'IE-Y jjZZFWԎ{;J4JQIQ,V&R` eKYJ%&X2\hOL88*q5=4VUGd" */lgcrvGs˜5(5I +p\p \"鍷!KGZ WNq.%e[ :$I3&AU i 5Ƅļi OS8(tNN﴾83oH>|d{?x;'o^;K,5ɨ':b:IC Ǖ)~DJUbV^Yx%F#2q}vqgYĹkJUzIFkD$Jɘ *0'3:d,01]2v>xǙ?0 7ׇ&`N06zQIS4Vs(SMFˬ&e5 -#VyP gIO 4$Jn S^F51RVTx?YNBw߿h-.:g@HNc4D8ZK [KT(9 (a BD]yvg], bm mi6 N 'AC *.2kڇBczaMxov3-Pv5qd0$RySc ĩ~y> g Aa%0T{/R) C!<gR&SŠtcGK[F$'ʗC$#, ]͟e7~^_H !--q[[rsKmͰflmE3oaҋ0{0bFӬNj6ɶk`7Vg/jZE_.SsQU^ҽ_ްqn-:`0WZ~A?nyk?]vS!; +=I´ZKWU8Sk^?hBw}]7뛋o{{Ax/޾q8.9\h~9JqLk^UY4mKhWYfpr ޯ-Twә暜JW- 1F0zaրVfYf ˚&4K/__X8v PjLxr*25^:`A&b K\;'^fǬm}lei0XO Yx>w:X߶;]wzU*ڙr28ÿpȺTBb$49)-K_ y1>Ƣ%w7e!4bqtJTTpD:dDm$`Y N%%=3ANy.Oq 0<0АILT19+-W^xA{ |Fb a4$TΑVđ6w;-8|O. >ɇB9͵؁%W3ǵHOi|ä' IZOi(nP4SHR c h†O&VsC g uI NY= Bp ,I+@Bpͭ@K\t݅?@o\|Gye{t(bkՆܫ F̻/6?:/JCrauøV#ߧKy>ni. ;ms+bB;ơލВ&gvm1mਣFg"@<8 "+(ɼEC].\2A(٩)T`{ ;)e`* tF\JDZGv$)&*BIPBg`H﫳?Tlnzm)|m9≮ /<{zwJ9~tiMDČjO-)vq?繒,8ä+ZV36s+XYSb`B.TIB2ǕnHocZ.?"^{:vPXb͟xΜ୍c!Dj9o|+|\oD zŸnbzWü% z,;\"t^o5 won¿웒0ov;gwM%O@˻,KŸWWnNk5#uBb1bZ7{5Ӛ^μhp:-ҭ_ Gz{;Ez/|U\WyM9[p=}:N`N.OCԉ?m'ؿ?qZzZigt() &dp>`! ^zV+y W/~??i ~7KG< miUN^YW7<Vq|X}qavzza5>{lfO{eh0a>e Ej [:x$#FaR BM"s,c7B>-%=mɠ yɠ+dTNxL< FdL6aoaOpv:̯ m%]| +2e2uC߲ʄ9r E^h o3J.iEUSIc vSuIKeysue,Cd/S,cgT?Y?8MҖ6ɎM=a_M;tŠ,HYZ#:Wi2l%y\-/!.2TؚlcĆg*8ǞG/ŏ6"ʤd50 8pV x&fr{jcdGM>'Q 8@&*6AĊ.QuDg瞌=Ac%\;ϛoorA?[4bvH1AF1Uknugnui.Л}mʭ>:խizS+6Xjs2pU{*pUpUtzW`y4 (n?|OWk0ZR?~pD٣ W;%e:ZX'Z}jցgj`O`-U5WSjv}j% p @[N`'OGVs \Uk5۾\2j66}Qt97nݕ֞r?djx BVwt5s 9w}smNΞ LWsѝ LF~oZ%,~?0m-\U5ל̚]UiwW(iO `NGWs_:DRwp~χjfN +jXG2]ټvK4C=Bq]jij!H(M*6Ȩ٘@:*8 '{F?zrjJ}?>Gbۉ;qD>8䱹@۱Zj3 WDՈ¤ΗddCƒ8;MO׼5WER#_Tz{fuuN{ g_osY Ëz|wק(ytm/osuwg>rox S:#]\.{UX뚛ܺ?J>],ܿ5eu\ů7K6h7u-u͗tCUWu )@bO,]n@sAdDfx _H랟ETG"]B1 ÏsBfBN&M1 rCrW@IbuY$Mj"&y <@b(եͭ/׸'~jCݾg+F+s{mz5K 4z{֬'0w46$foWĭL"]uZ>oV6KceP7Ujȋ!o Nc_ o'LHQ Uovsy/]pYК5"kYL@/c ʢPPLwFQGQ8T|ޑv,pI$EI9# F*5*1/! ~6̜Pt|mlG`ۊTZ-izV@7[K~m݋aNL12TU0 ;ej1IXƘVXh%9Q\BB*h(!PG%90z4f9×}e~kzO~ml.X h92ui t~zAc0@qcc٢t74a7ִ&;-հzb[gۏZ^Z9/ lX̀E N$Cv¥T*-\1APRRvR5Q@HH^͆ażd&"Ă6 ұ&C)2瀝# p*'ILo3/fqެ,\v[ѣn79ؑ'q 6±XS8YJCwq\gxs,3fb j܅ UZ&"*V{A=Gӵ.o=Mzy?'iyNpK{|MAQp뵼Ń)^6^SAXn=nyt7bh<4nd*aOyht%/qm[ӯ揓ͤ%KOygOHػp94><΍-b _ك'snz]{P[+#{oT:(7R(@_Z6oK50'lp"+B|P(wM^kQܽ{EqS%AD^6%XAb65/8SsAچ$}B}Z?v>x3yu@s~'sR?= 獨i4eN qss@Ǯ-Eߘl ̪K 5 |:ѹ sAwM\PKxqfp+Bk:(D@! ykA H:twMq kzYy.RrŻ5FRe)T!0,*#r%g{tM]h"+c@gӀ<d"CΒ^@M:1gyW{ӃOZۅ&*/n/k"IPFڀ*( )EQ)Zz9t7; V}!Y3jdA&+(XT!@IP TBX'{k)+ f3 1,e0tD&10,fsȀa'їܒWspB];Msu@eOhoFY܎\X)`A:nE|凮v\,>]gel6Lg~ ƨ7)u(T.骀[N|NO~xwoϏ?ӻ~qiA ;Lݔ8:>NO'͓,뺡fܜSZt u]9eiom7;Nݿߗso|1L `k}x0}zEuަs㚷*ZE+oq[Fxg[F8!"k8Ѡ ^HA)d TK fek=x O?̵M# ms0VQDϔP;u땂O9c/&QXuIN#'90KĆ2Q27=<1$K7O}xH5a]%H[)4&`y%%2Nz 6RBRVyQފnU؀:Y4s);"^5`X TGfl0q֘s)Ԃ֞Q=ޥa[t/vƠyb'cLuH'*+6jIAP GG$Jڬ/aR/%G.Y+[y)#"b(j FрG!eqcy5`~Z,JhoFqݱh;.(9s2 Yb2x~ܗOg/ kq28Hg7)A]=FD-Qvi.WӸRHU"s ˬ KF B3daۍF6Һ#|f7.NnL] -(j?F"xW <yEki's%ƂDIQ {ƴ-4<ҷS =-|o8)V$vFh-qQ:5 4*K!75iK[jړU%Yk{HU۹y |eނ]"QeifźuΗ]O?Uؤ X${ PUj$\Q)oUc>,U$|]Z,xiv'cƖNB4ynl92R:,_x^+ q 6XĠi!! = jX v `{(3)nB^p.RNC'S Ec.Fk@Ni,K#F+ ȥHwS-,@/@/ _0|ѬWዪ6Rn=Vq%[g(%Ω339#3!˨; 4 d>90̫0ʿxMHF`RGSJmsm|`w1z&oisuFn '[6nn|vz C!fH6)"ߜ <Xek cש\ĉ&>_ːMF ~ې]ϗ\Z ?ÊCAl0 w/u|7uUft_`.v;3e홝lܹ__&w6_3d|o+",sq_.-z&UԞ0< yӉnI€n\oP (+)f@QL1Swf6P`?vksN\}H߷'l'-߱c:O0($deg1'΄w G%:zU%AZX:,~YlVE9Y}r](y2,ƫ8NܙCg9|d׳qRR_gg=cAL<J2q(pb*IY W/X+|0pĥP*I+ԾUR^ \Q0W 0`*KLJpTWLR \C$-WIJ]DXp69ES@j*W4{=sP+= !~)"/W٠7*zkx9wxIAVuJ)X !|{~fvoG뢣t}&9h&Uc 9K19uöwP1DM0. &cA:n/IJ6pF*VD+;kHE>ג)E 0÷zɚ%77u()?$K.9^'/ƒR!,ݒKR2Zr/В @`UC$-c$-\@BəED< ӴGɱl.dpbjyduL\9FZu0 >6R"4Ev*4EYqyOX~Ooˬ3v/~"(q*9;6 \>G,OD:{I1.ѧ0'xo:0ޘn3~P5z,M7Mͤz7̐R %̘zh#X6ꨌYz`Ifx -b0fQj|w1a,::m㽊Tr󩧨TW]6HL OҊ߫HR $i|8 ޻ |xu|//`ZA6xg1|nIWM~8U{rG&WHk&*8[z#zIZz#+zI^RjGG$j=[K3ʰ|;FJepv+*Y}hӢWuj WW^퇯QL!oty`iיn\=Գ|EUi-u2,VMqEWڬ&|/yUk?Ep3[k?mN ѐm^U /ܯI39Q}"0ҵeH\Qڢ rL^Yˌ(HFtFFfntGF:ݏ9UaVI`GC JJ%Arlx hreqctK-m=3{ Fs46l7&j#}2f?&~3~g/'>يeZSGhjJy3=Q>_H84WZͫ Oj>+xX9"kla6zK~<?$o#Y-|gFi V&OJ5f < C`g(S A?{Wƍ`,p/&9|' Cn..M Bd+qp$ٲilC*F$g9UL2&Y:Ftt>io0:E:ZD]|&]5g;) x8Vw!J"oi ғ&e|zX*6mWErh,&#CEDI"A";̣"E)d!l"x:D,WP8.H&E.-9KB( rXa Ɏm츐zr<¹"EPQDzҞU@E 9J3APKS M"T B`-9"xel9+Ș9im >L`58&UbqrojF]6vL6mBa%1$xH* ]hƎ&pr|նMNhkܯm`P|p6ÁN>Z cٙkr@aB)sAK%B#;wܧb?Ws!S93KLJ,,9(dGiTY^XXugYPCa`^`HQlur*0C>6t9;#Ox }ڐ|L|Ȼ˷ yiml2퇏w2Ʃ䱫V^[Tsc27Nk?$I-#uցG&EB|5dHz#%Qs7 Y%ge^m X }0s7Z3tkcoy$r|F})R-w,1%*SDuF1 ؘ K;ݦ5I0q& ."aϊO =, d$rzA` uZSWnnҰxğ(*`KYeZu+[YɽUfep“ bv'Қo1)x&J)$prve7X Z\JL[=]mMN)qM<* z0J{'fTfZxƔ@kd9;lt}Pt[:ųŒ_oUqrop8_[dGp tsjE2JP9ِ euaE&jLdɠIt(3Dc^9;Gi<lt3gysZ复:8&ER@|ĀdJYɛB0ɱS5I'(fG+Uv %74*Pq"<*>rUt}t)RU)[Yg52CtLZU F߸5ⳏ#-rDI79q,;$$%Ȁg'J-cD(K$>mQۮ{3muYhuFn B,\,^fK-E7z2Uo ;}=΅'G8۫׮7]3/1WRLRZ'Bbi,ART̳~??>]t뢓k(,̔=YZ-:FRe7 i"q(4a@(54.XQ0Kܰ3$Rrv-n7;JKyU5_Wg%A`RBgBuO.gg W'Vc\ Fccչ8٬τ{[kߪZƣ_.g3?ZR qhI0ZTk:oW#Uk5LTgUѸ?[,(; ?-_`lZVk: 7NK\OکbE$Gƹ*QB l( ^wSo_.?˨x+kij~#*-R;O-iE4~FG_WqpqFF~~sW?9eܞקo~zE8W".wp~=wp}T[VM@Kծ˧^UGn9Vprs3=0VMU'nk*SCljQcV0[ nW:&חNN#;- qL89m]ڲ,1 y9_`ZЍ>9Wi2͡uwF?mno<5"< 7:QC/G)mu%+^Vvψ==#֙,{@5Og@ksFz)1™ ;İz>cȷG4S !)@l9ImH1dZ,̌gdv+9Lqvyq,iK=Ӗ㴥}WY:5ynӪ>Yiu|0[rC;BmAZ9KNpCЩ^?1%˸C:x쭞 9TƬ4kYk/ Al XNw4ݐVߡ}cE?  mk=A@/hdG| /~E=8a73^9Q`|NDLD^5QЉi[rlQxaa|C+yFadG*?)(eF+0% *%hLbTLF~]x 0?-¾N8h4WQZ1wzT{w2酏Xᇒᴸb upqPj:ܺ:8طկ嗘:~h?q0FVEpiw*%v{;%]^V{w)N_q I&۸!{S<ߥxv5oDo. Vl<(R(] MVe -`#ܢi%cr~1R.[kț.ߚUb_TU>b ע %I3='oe;(tq3z)aôg̟gٓ =!wWUV3ak8NXE)3ɍPr*X򚮀:aСoPfl\8 깿_\j` _,hԭ4j\3QS}6xwV~0a0>54\"K-( ?8w+B>\ @c7AmtAh,KIN cUIdLYC\E:-aWOm]7s~waf?5VD-Ls?v9Gg9k@"~ R$X[LqwtWjy?[|'O;}>;&~l_NN?9 &7~8;?I7O'"L_K'Ip֗pE;yėcXI͏oy; ۷5V0ksy 9=9Y£큠fDm &+̖!Im3gUm _+楹-q,19M{DNIb?)=1ՀN1.?A5/f ,i$Nk(js7 }.H 0u? ӶIj˽MAVfPh-޶t`mܰj &'iܬy)>a>MDQ>fs{Z$BO`LwcBF@lMc(:ʌrP|YB#//3SD$@7\ǘ|h)eɫf8bxrI`{.E:*#XZT 㨜YЗ@NՔ Dk< :KY˽ZZ gӷxi 5<\Hv%L3$ftcuFۼk9FL3PviRjůt~9KO0NS)3fx qݔZh\;`=}ع\f&tn*Iժgԓ-Fb!;4l3ަĤkV_&9kS+UR[1Э+m|uaw1YKwf`{i]tG׾r'P0vY4a#Mx>9HysQ b+M5^Lg>z<>f[ &Qw| % ujkumxɵ.ۉ_ =~_z{Xq4H^131mSu^o70g5g70tn] C}Nh?@’(7M x S-Nk.U>Z)nq]]KqA10 fbrH8+CA|YZ]z W!=]VQI ЫXL8P=tHbJ苮G׺q/@ZzOzYó6NdcQ6mU@Iޕq`g7K׽ W~ջ|td؞w%?RvB{29Kl FQϕD,#aA6k]=lG$ܵbh]BW5[tm?]50) rP܈KSfڑTSuMig`1V1ZlBJlrD`qɼy߫gu< &qũ{v}?W)jCs]Q,Y KJ%P%P%P%U=a%P%P%P%P%P%P%P,-KB;qs`p? ffɸlftԿ}Vljv2IoHeءW?:԰fޮO Jq}ŰӆU6XZM%OUZL!*s\@˻&"yX=W݃5w?-坼SIgo fǦl⭧r~s {emgҗ̝v/.MP""PP@ (BqEdvEHAήi>hViB`-z$JeC>KAŪV Y* 1.+JU%}lr'AUT°`+9f*uh%r4,dXpY`sa&x92JEdY (t)$176w6r6@BW SgMq:vU^HSD€MQT`wk7d2PpOx tr:ETl 'uyG5sGG0n]Uek]wGqMX3k"Qnz,ϜrYy LCw햼KM);DZ  0r1;l-QKaj<ʹܣCPZ*GH!,8d(kX:9U/zi2~JJ5Π8L$Nΐ/T\YRޯ8'QၱXiT TV0bRs8.z/#6iPFZ"83HayblHzoT@C[miDVގ͚ղ@:"eGOTаާ8N-IX+s^A簌|hO.Wmvt tњ3 &cUѰCI^r$Q+-9R'MEIzp /{<_| )c# %%iL3CX@*gh#.(VBL_Wּ%Ԟ&@f*B><Ryb5#AνcL`(eP1% r!2dJFHqb'8> hV12XP1g&-L6]%,BY" * gMf/Y7ߌ]{4L7'/\bӈ#@Hc$X#r`Q2c ͵$WX?E#KO0'ZK(Ib yI:U$LEt!"'U_.r6Kl;9Xv6(RhBZmdZ[f 7I#G$2L" Ε `J1#!<gA%b`!f< A = =r`TiLfy6 I0DJDY""D\"QǼ :uk,HF"@FbrykWj0qbKFb3[XA`Ir,i$,N.7je6r6K/L8uLl\\"/ru1gbf K8P4 b@sfNqx,0w<1 x"lV7Iȍgƕ4:x$n|tcEϨbN y6?UZbUr.źTq*V${NCrǞJ8!+ƑҮ! *8KVjOǔM(+EV6&HAI~cqM8SVaeXd[mP8P"рeΝ}LX=]Lvu ou>-kIe@TC]^/5\L&кSi{xNgmDz]mB ZNw[ceZ`FnǼ oǶ<7t {/zM-kN?mm͖B߶yJ/`9 L5_- .kVLQ |<[v#K2c j4l$ @w0јv~/x :t@tB@h-U:B})ߤs: _kTW~#D)r9Mۜw ? 9/{ō؆-M|`ȵIKz&l5x\zpWI#9[n, PIXfRT(~ȦH)Rs+5y{~dͭ7. =lɏab:#Y}RGfYQf54AIM%G̈2t0SR;t( *~y}y!shxzmᰳDy~ލG jo/U$/Wyk=c6{u)cW* ]P U ytrທ}hǨ~w0h 0 any*l}[\n0Q/0H[+aK)U4"NYΘs*#W}8<9MߟFqY~s}x:Ln>9=Lk3 =*<*>&E=JkQ'fR>1o4PWG+uv:敖WAdyLxA?00[3}-}?܌\/C0`Ȥ̾^v_ӛ\\fѧ7fwH;<{ۻ,O+0Dܛy/up>}~u>)~}Sb X;鱀VF}X?.ڛZj9c"RWWY^)'jTio}Ի^+c0' p3\A@njډx#iԖk5[:eVV{N늏LRp~F60כ5t>j\I329I{v S;:\6|{>y3?x8 V29jh!! =[P5'V#e$z(39S[X'b?[ h!/ x?V2O6<1&)=)eɫf8bxrI`{.E:*#X@Zz],ar3ܙMC-_6OѢ^ gc(s s^|GA=Nײlۜ'3%Xdi %.L}PfRvnjoS P6K "Fc)$@)C1ŬL# 6W!B'35&m@:U&LI8FҎYA<@Y/5^#-έ4!͕}JXRyׯ>-^ (n%o%u/r2e9 ,:3xb `kBZkߘ!)#5 ǀ%!/:<(]xgΓS")A 8&NF03,*ȁtfD00`QiH&>nq:Q P ғuהj*OASV}0)S XXE^G[G ޳amsYʢAs#U|h E`حeѳ!c^-%pT'Riԁ*>)*P(ϩלImMcDDbgb6_^̗y7E?Qn`W (>kv!\o&WŅs9#7H pqmu׽ngq kIyre*E/10Q,11uҌF4XќICzD7~N;z<ٽNj\\+Ac6 kc[@s ulf~0r 4ب Kwp1+yff;;vi?v~|Ĥ㢺$$ǝO_)p5Lկ _&vUq+$߼w/_Ϟ_ˋ/.0Q'ߞ\|rIp JI[Vm E[5uP6[]wx9i x?M􂜬%B%,Xt47\jVn/d?Ύ,%ìZIz6z]xtөr(]"8:m>,:ٻee jA^7:bZn&: ]gA4mЄ[n~FZҪgR@%e`}ZUU"ٺI H'D p4(h"$iPc-S/)9麗0+v1L2Ϟ?wͰ!FNA QXE=SBYlAcW rl (,:NN+;9;ZOlZ+FhѦQ|it YVuHEYͤ::/~u46TzQS:!3IúԦR9~(%^ +`r6)gm]$R%aƟrF N/TEʳtݤ1\F[xt|9M'-AَwQ+s~Ry\\ϔ D\f"ˎ/3>{>Gth:XAR#Ml0YC*SW](j[-PU'ϺpهP6]-dɏϲ:RT;ui۟"Դbxgr3ꏫ^Nj'H.5 g ,nq>zg /"B$ɩ6)NRW>|n=7"y1ƽ  !)6 `ڰ ,IƌQ@2n  {LZ 9 4BFm9:K8 P׾5:A:EvTr-NQY4C J׀uIYL|YRo|cb,f\dDRRUrb/fw#{OI\'#2ꩈD-û.|{%W%R+.YyS,{gQ ) K8 :OvX Lq;q-qF'$=1ȕDc;̶b/:bZb~BJb~2*SWZFpR++)TA${k] (%Mt\|8yf.;tWi%ĕ ^riNjA,m >h2Ͽ[{SrP, <;6'PՌU r0_ySHb)sXtsb{YKy*Ȍ~ 1v )>Nw]WEmnQ7~IXO)zw!N,g0t_ԋ@SeiIOʴ}RrqJ/[%%ߙ| 3꼇2x!XF ,Z1$Z*o#J]@:j)(aDDa)39B `pDX11C-[ۦ[gqh)̓X~(+4t]nM/>뱿1Ej:|ِۄ 3G`%h3ʍy^e1qj\ 7hS2$ X#5n[iC޾c|/˝kv˫|'ږGQVbzkY;ݽ9Max1֋b[3?@˷ >OQ~9fH>&qn߻~Q)Kg[b"P\E^,úy1Pkhp9yz[IC_y8h /l:Jfp;W%Y%8 RT2Z&@UVO5%I;)͎|;rcyy (9|1z"\0H(s6(b^dT8_nKBr}_ݷ~c7zFP.%U;3b<`m|5w:m{wxe%Qspyל뜅/eVJּ3.J{<@N9'Tm N1`%*1˫ oUŌ*U$9 (-sGb=UIMH; .F(b/ަȸWBKO!\TU$. u(|q-`"fp,i"0O}DSɍAg&6b7x9Y IhRЛΩ`(&$?R"22<=&D eDsZ.$G:jxXLxX;F ]Abq,"ˆ(zDqƫ,3^9 s`Ra֡0G։0G!jY m߄2)j@$tuR9IHK9v)v~Ե3sfu3 u9~+NYj :zmP Sx7(a [[Wֻ9tf1O[Vز;]oIs@k-wl2_{vGƎ?@ 0Ϧ='㵄w'.sٱاMww| _myJ:,.{3ʵ|"y",x2]"IE\Jpڑ*L8'@y;'ry"9CN:n c OMԚPa5J nF hma6jh- 0~ TzŲ*QǸ `] dp1V2V)q; .wA?2rq~O&!% F۔OISdOS ^-4ujILW;3iܩ?rZ~,p^V'c9 \v{JGtI]#yC 'Xζ;8GNjA+ELԓȑg!D.3F %ÈcNOEKu h!(Жktp-b4H)GP1%9`iSd: zE0Q< `L%*FF)S^@<>Sx&Ug+!Djs9\ڒԃBpy&2I ܺ¾ϋ<e\ !>8fxVl:+ERRJ'Nhy<HJ,[ %Nƅwkcݤ[lPQ8d2OҞM7xR 4 V"/(vo^?x8j?-G?WN/6PS}؈DfJW|p4\8&z+ŗkWpYWnx9ƛq7ú?P 3hջ\_-c6{L&aZP'AQ1BeJY}L0 _=J.~4wKAoo'ͱo={;,&~2xY>y.{]E720xhYƫ_Q; /"-i|a[VyWW54^Ν2|Cxy~lٟ'#x՛XwW?og.)*.;hq*뵛~ɍhЎ_BA^:Q?7o >G7MxW"} (s~k~ph4Q7~5hyh.}Ϝj>p.LVUmzy2?|ey[֔׫1l-O9AŅ:qzt8n6}qK<}uW͕rn!ߎu- _ X׭1M3_y2%:4rPN-NR{q?4?we'T`W'q5Cr$l]8UlAEj2<Pob oc-/դ]XSvZeͱBgji\;r";=^v/= ]C!@ХQy%$P9K{.q"E/B2YxN&{寏A(e91WHSKyrb3yn3KPBZjEO_-,zY}qiڙ̥t̤puf]mo+B/g M{n{[M 0"KdI/]I_7ɶ#R$ >sJ3tP*I9H)A? =IDZG7j&81H%iDwt~^-$E2 ,oG娜 !)`lcB+G{r6tٽeZ/i$v✑oQx*KZJ2 8FeT5F+PڶoYb) kPXK !*h.\+E9zڔu~ӰTy; \w !˟)0ni^hbLkȪGsPX&WBs {IR,%|XOror{B`\ŸJ4 6NдrR9+c{E+l)óC=ReAfJ'')>H-*I2VeLX}F{Oh}O푭'W\qe,h%I蔴΋l]?ww޴78]^xqtw!s7Qiߝ^Xh:pYl&8}ҒvkKJٶff 7WͼŋQ]?4eM'Wjs|6pZmmUݭr[[ juQܡx |- kZ>cl,G_zgj[6knoNj.Ys{|6a4Jhfpem~+)w<(OIH=o+~s9滷'\oHO#[Gw:[M>iSijo4)COբ#oi ==p+~]sUs]T͙??nJ&4v9Z@ijVgij}ݪif|eT_~hL&`f<{ H Is-1)2R#J-;ګ`] -i>zb 34O)aI3erƅC4}HUT3E6Ae|xSMNB{mZt߷QǼ2]=:9uX5YJɤgnyyH Fx`ux:_p%UN:.9;]#t,ˌ~8axa1Y^h4hoB"K.IDsc,SeTCU u!"pnV} thAZKIXd1RI:e#dR{@3)[P)iB \(cbY`/mF)_fG}:HpbV៻7>`E}ni|v_?i3OuUz~TΘM7)ـц+͓Gʗ.Kwt#_/rx NTb\36zDrXafuO'ZDH-W|ᠸ )+\deNܗ&҃a*j9ЗSe-Փ u$tNd֘J)S+h0Yp1Y@eU  I@3$}ּ|JEU !dTJȘ :kZw *I9Ra<_X&w;lҶP|uYd7o%JkMӭUK&Z㕷.&nUeՠI{8:: [DfY!m,&njFSR0rC8*K.H`,Dm*IUS9gRR5c׌J5]XmeWʺPЯS_\d~w}rq#w~WxqklyP<0x8Y GJФ*.dTqS7t2+b$ AHQԦ46JV@ٷ* 䈙Ej՝;̦i֮:= B6K c0dْ2iFx2 ZI2vF5JP ) O>FtMB@A}R։E2pպ_F2(ƾhjQUֈjЈFܹM&l1h (pL1jC)Zi[M3H30E1j%9dcLV\LH!: ]dI eI8yɯksKWkHSH/.pYmW/VfzЋ^yp" 4Y}I %wHZ20D26Hp HQpR2PiS^|}XmuE3&TXtw]]v<(s_'@0 !ҭWH?ṫخ=/96n$y8b̍ޑy4W*zh-)w.bscz<8fZn+\J+f-\%nL*vno;A%@FI:04M$$BPtLFi;#]H!1FQ;q3˓dI 셰u @E]R qWNhi!qt LAH<)D'Z0LFkug}2}qv~sݼb+v$[Aa1aĭ PzyojF8It4䘙1 @ BV69.BNjKmPC8Xdl+_}~?w ܁OBkx(7/4Kh"WJ5*5J1t Fi4ZRl(걲Ά-ֳLMd(`^YhM)# 11gښ۸_=pU(UYWlrT*\%!CRW6f7IC݆)[ 4FF*Cz)U .7_~Qn;tLH JБX~`vZ\0s,G6@y$"kY"4aw 1 A\pU_a] y?=΋r١QR(|yfrcVҰ|[ƚpQS,ʛ̏Ӑ-p<@! i()K,ol0ȥ^4M}57E%MǪ|v&\˹ c\X1˨3bVr RLen̚3Ԓ骚[6nzH#o 85ql-o3k<3"Yn֊ΦPAPLfb˽])[twf 169[$h,Db,bz0J!D̔ט΃=4q~ôhaӹ@!hd ; E8jQH*1CSIDxPRX ţՃwH  vd)SP@:2O+@2AZpx+q8 E"Vu^7/ֻT#,a^`b1sN)M}@3±{呂ޱWoZ4oT<1[ˢg12 BB%3Q-Cb:lIb@ I)ӭ,EFxNLjh$";CC;:3/d8 Le?dEsXi2NX~Txw>=,o.m}Tqe4>]YV^bXr>-xxjhb7t infX4sŰ4I@Omu㬱Uַ:dS}-Z$3e8ϤSإAG͆1KV}gʎycd1%.(]O*Y@ÑssVݘ pg W2HڟdTQ8z?>|xާOc?8>ȩLKFFIQnj{4-jڛ5 .M6{].7{uid-*ye{e3-n]4W;3cPR$ͮ(fJ(hֹ3^ ^z'I[TWʄ^o<2jI7ǕbZWTuƝ9f i?fMmǶ8?^"|KR@'i~ʰ=/|YӡttZZ֣foֽrH'D`mG84ki4(ű1"%zAYiFOmbmYyzF3JumzиB$` y*zg"cR3\^0²#NrZ9əh[2FZӨǜ缌bNϯªYyuXmyeRo|my'Dh6WH8ȓN.0"~ ﹻ1j rBxXr@? ֹvC`q; y Jۉ)8%b"XZYh6 Z#v2q<.W_ +4ċ/|wm5R{\8_{Z᳧5!j-z=B4FG ;nИz|t0lTc䕔8u$H) K!6GXޔ~-N<r-|7Ϥ0١<`Aw&,[BD ۞ty`_MXL%us lK{[0X +A1؜arJnέ$,<ՙ_6pmyu#.c7}Ơ}KN0ǒ븑N8MU Vl#b,*HݫgK*BX-KM'b FрG!e@-+sǃ5.k)t1 q4[2mE9RƢO _#yt=aB#n P?Mr@j`QpK㴓k6C'A0`>5Y AZ?x%%HS2ఌݿa$ߜ6:M ywzP͵JܞniN2vVnWqnXd&a)(A5 l$@R\;XhL; vp<=i=yiqĥ(`I$vFh-qQ:5 { Os+aF 1G~>a[ ։|~?9_vA={lxcSsEn,y^]دߝWX7?"Jk0mW~Z )C0-jXL%Θ ]:`"2fZNw8 wX~q8Nzo˿+F)q<ptq/ʻɌO´ e e%o{7Ϳ4 9^ -1o2P{!XoR{BoneJc42Ic.Ehu<,HTy1 @8;OtTI7oRvt掽MՏaٲedjR+?9k-[$ ʼTO%N9;ACkaEn +rݓ~Y#"\t,L~>1(ʚ̢= .46ӟ@ƸEcaWKl7/U_L~"baR IUe5ժgm}PRjtqK- n֯& Bn,Bեngk-A[ט|'IYER뚆f{`YŽy]ʒbhÒYjjYW_w\ VRD)=1D@G j DOeF9M,.'[!D_PKd-!<$y=Fk@Ni,K&K#gēm3xϥH:2 VwZ؎KIUZ~[rF跺Exhn@uDTwEQJۓ@ȁi3$ꞖFkFIj:Rjs1R"b"armɑ a"Vtc=vj3P~(G{\E`)Ӳvp$uh%r@iQ dXpbO9NMG EέR'VYqGT G)𗤐k49cO~!c\knTN;KyϧՙuA灘!*SJ 䣁MC2 ֎ U{:3mXnyTd8 sQed(@4b}J]@:)9X:g:r8"䔦KW~|Q|nV;cZ`hc@Q=0-Of<ˈZ,Zؙp( Z uUp}_QVN~zK>J8'rҒ'%ߙ\w%-9P+0J"0-9%+ r3*+ĮDF.Ju+F1!tU"XW\-wE\txqzJUMSO0_ bg~lR?$0F3oyrF代ǐ*OD'E%fSx;(4c9/Jy/,30ٰׯݡ wk=rh8R5LQ2~6aOe>'3>m 5׾P~ᇷ)O!3Cyp(A=BljY~3Sf2{Ǽ=8y00D*r 'mX >l&+DNO ZO*ՠ L꼛DB+`xY^vG(6TLb_\/쇣=y}wo^pr:;kS}2ւ`۬y=M MFкYW Ƹ49qoҟ&zOV7,iv2Op:Ӯ ň_ZL+IY&\/YOsg:\XQa TyNpH+xB3$UԥPs9+LXLY0]g6gh[n;vM)"j͗@ wmme`\Zъ UԊs^ީfD7iㄈ XMD= Jq%3KI,®DYjF}09zk="!#JRhL@=>: 2lTc䕔8u$H) !Lj⭤}-nP5'<\ډoYWiA](Dmo9iUO^Y:|N |@!ᴃhd;:Ȇq;<0ޡPS&T̾ sevEm&&CMlG>dem6ODSv]ƈm9Yj:RjsfSb< &zG0#b3ɹ`1X>^V+n1v[mSʼncg-@rFϵ,wJÃ&-79s> 2BɤdHTR2$hw ۧdHP'oS2|3)8B-_hёBFSCP#3 `rAĤ;i6ondL(_dU~&n! ~Oj 7RZȣ)IHTbZ-0bQ]_(#%m]/}݄73zL׵m #ǫje/'ϏW,G2A1 ղe#s'O@QL #PBHƛi) XIJ4i 9G4` ɑ4s#M:mPу JyԲMR4O ک71MKuQeuyM,F0U LjSK|ČHIhƈ"HO`z P$F~7=<4W0n}8a!¾>_=f;SWyb^^M%OQVoӄ`QZ1=O~8H9D3=2 Cw O ܟM<L+`$A>w_%޹I_0kkS)t[&XQ=>W +lu8|~{)`i~~FHomO'j{5A^-{]$]+73V.߿ퟧu^ؗ闄)GS1hT}h,յn{,Q8frHVZcocaAF)hpp<89K_yxVuV44T;cܣ+坦Y1&i諟A"F;)MŒ?\J՜w]QzZwhbtW,ׯ; vS5 Twtmz[|G%QER{&{5`Žqk9>MF{SM|ƒR:%Che 4l4 wcBF@yc(:ʌr޸4rq>3!<25`~g91,]YA?#6AQ0="9T0hZ0|6fŝ9k"+nFɯ Wq_W(dWWKD^ֱ"ocw67}>Rp4[/'FPד ~Y^G3nWx?+UGbr#fTtFO.B\5^+k)L@yM>jb wYbYc}*=L5`}e8[QqD=y'.Rhiϫn@Rx4ɉ$gL\sPR3C+} SF+#?`i>AvB{29Kl FQϕD,#偠Atbw{kNٲWFǡRZ -8E|Ⱥ?]q;z8@8+RV䌡 A[#: qѺQǴRӘGRGO8XF(zEY>fIH㚥^kЃEG7aAT"HKsʵ&G#Xjj{=@9IBsF)c#)XG(J#J ÂkT4qTQ(uvEw@pRX IJNcɍs :_DL׊ K<Ւ}h'yي!~Z6G+b> \Ӈ\%̯z4 ɴئ/?f\TRF9ӆ6XM%O'!.*&-2w!dЍ`i^. ủډ S,39BBxV`PrJѥ|n$~ޯa+r | 3꼇\e׌`^ %P ' %Rm 5VU3 o(%gMk ojꮏ3s<+EؤJ@9Чq /_SSW/R=7ftrJtяr0@Sa\-p}~ c$ `;(ݷ1E@ Kd688 :+p*/2S/*  opFL2 .mKrx"O`f^p vyuGŷ_^Yyzs\:OlmVBwd ;mVe,]-gv|TMMsѿ<σT~>_Nr狯o=]%+]nIX`v1]r^"/a_ `evڂiոYxgQ eTu2dDdf16H#A seY.l DڅVrE oR+-IIQFDYHr^D([ndʣW$dc 4\JZea˾F)E\ G4+CB77DڼTiᅪnsw_ꎿUz,zmK7#/3̩^z6{Y"4Y[ߍ=h++ Nh3)ׂ.b'7\B I ru5A٥Ô?{mww~ ŝWme+-/Tj͓׆{n<߉»ѕVWl/.ISjsmCuO68|铍j֧J4B*xl^|*zf @"9HIHgIWш}7xhvه/>xѐwmwUB& t`cD~IfXS@ϋj+z3;!S)Z_72WݜouWoݫT/{+xp p)[( 2i 5Η=Qk:O]+C4uVp{6ښlnl4膯ieO4ʿJYz4D/`@B`/pkuwi-W Ux )'b*D>;yKfiwVemOV$,SѓU R"Y\9 L̈́. g+Xwsk&H# M^_= S))g*5?Q81ކo+ۢӧZٖYjiO?܇mK[| 9$C26E. 'M) 2# % Rkφk/(+L zdBRʃz ;묗R%Ph8h4sDw^tAҞ"hȧMd@Cٚ :Qs˜V>)2\jy9Ar6+J,mP5k#L)cK&y0͹OMYaMv.Heq.'0ɷQXƘ i=H2 F)h1[I1f}&cEc7ݷ3X{ tB6oHY?4IԦɖb?JռP)i5N%\`ɺkP`U2~9d<\-y,aߔD< (OYZd^)' \LpZ>͍LiiقW._T*U[--F4>>~mNlZc];vvHAE Zdm/,B(E`ar V߯ ϳ/JiޏMt"WY$QdB PS}L}HMl)b]X`]Ya;n68l4JJfBB~v<-6ܐK`8RqVJz,w^T,E9yyy*e#NNTK /<9X,TP J],Ԧw6igë YJ;zߟ;p~Jy͓`\L!^L Zilۤ/A9EiUE0J!+9(.۬^+M (c8˨zHYE(*6 `XLp"wFE5{^T3% / N}b`*XY1@Q4Nڑ-ɜH:E96D#*nkf{id 8Cϊbr" )Ly1̆55g@Mk/R򟟕`>c|n oǿ%٦^EoI.t U>wBFYyP{FyI}dxX8CU$X#8zI7[_[cZ, I3<5qe"18nWi SP6̅‹rᚰx[g_X}m[n7_[fUb"%3L£Hf+hPBL:z9h}&klv5 {. &Rڔ ^C4ɮL:Y0Ոgl7+!f[X1Y`;[ly7#Q1G sFRV fDj39?.xr"d K&? qML,ha$8>p>fY::_bl #6f2"4̈3bψ;-^g) zy2 #rt j? C$gaȷ1ke4 r}tdCHZ!.9QW{BīMƬT^ y]p(sVRT(Z1KϴQ,KHDPUf =/^‡YG[=N; ])~G ;ts9Op]9Oh%m?_PJG42!* ]Fw Z4m+qwCWp`Sp h8LWG KWǡx:vAWթ]ϕ/:DWXI*pv Zm+B9-ҕAם"(Z\rWw&7;7Lwo5\ۨ;{ r`݉, \+j]52CtE5U;YR J=]AȺ];ȮUADhƸ ҕVӶguҿ ֣)Wsɢ+ kNQ 0Aď˼HosH3Uqd;` o(N4yۭa]* ` `WhKgΧR Mk%XUkDW誠֑( oFթ 3tEp5]V J-ҕ5^ssޝ%Up΢[@(WD"]ѶSG0 *puv\NW~?tzu Z}Q+qp8J'C)[ #Jtujs+d+l ]BW-UAYOWo '>ɤ!{y|r8Π &F ]`v<+">tx4gN?~|(HbFAN%Y+|JE8bAbfi*'(jU+ +<"#I,.l](咔̺)K۔^*)\ r ճzXm߳'!Coٝ-ۭ,޶=7 ʪr# VAfqjQdyҰم :dg+wP)!l0CiIg9{7JSGwO ^q)n$ 9I4O(`ɺh@$b&{<_(BiR{sUvġ.#뗻gdA;YѬZ?nx%P=v^F~&bK!jg/2!"]{8A+T!m4ҲMo|6g\L/u"0eB+s`X'L^,q^ }DP"FUIk9 |01x9 GfD1@VGT 6gsB:Svpy3'(SjwE r+ gzft `:}|۪슲ɱۥ;;Ҍ[~1TN!9I:I[}ztj=yˋσQ=wfF]OvqExk\!wi._m~6]5b.&p7Vl ݧ~<~N+#L`k}.g}qE `m _8p-GG#f+֘ʣ杤ÆKZ:u_je]svNKu%>]-pzxZmR!t_p>7@ |ׁyށ:ǁ2e2e2eRJ=,urJX /<9 b{hI*Ņ'.E% \g*/Gl{Cц,25g9 )ˤO2mmΆWtwV$v&?rΪR=+%6Os%2Cx 3x \ ][Ъg/( w]ӑ|uřqaE 7T%(@9(B)d4ţU>XΧe4LHqBxuhS¿"P@z-,DZڡUD 1ؔhK`95{m*lf8֤R'"Do9҃=pcp#H6#䢱x\ɞR;GuveKCbRҮpH2, %ʞg@LϠF7Rr& | T%ڣe?5qSȳmT?ren&?K%p`~iц4}/'4T"Sr'%^r<]{=x:#1,ؘ $ǝa5IfgNdE$[B#DptP)ll;-q*I ?pt@\jQdiZJZu4!׬nۃ..72j()zg6ON^}km͹-=c8V.3rv 2:1lTJ!1 :oxsgo k v_"b.hŚHiEC\4ں\L9G0DzTR1_Ybcrc96ptHaA9 hHs:}%&We&71\f'Lv!3¡,?.u,ܵyny u{VȍM#^骑/Dhsץ׏vĠ1~?%u}u5m{z-lv2iRˊZvv;ojB077;&Tݦ;,gܭӽL{.s Usv;ߖMoHcb?>e=7"ߜn?LMWi1ͫzlqF |GJ :0U!H#tNtq{9“tqN1F+cGh iYgp CXmC8Ep l /!tJޅ$T)Df'q3LZgS\>$F3p?d5_q&Kb(X& J^\f):v ^_/98߉\^ϝR̘9r9BTh\1,(da>2皇,CTFx[1ކ{<Ț zM7_yEoq?N{eʦ%:x 1 Y+2uLK>X z F%:D "]nԸ fcsT)'tc0di`&wiL]:XWwڗ!;$jVVc؞bD%@sǙ|U`ɺT`DR=3:F-c DRU#(RdhO;3_zBO<DŔѦT <-i)(pA\9 &p<߲/"O>ޟ-u 8^=1r<{'C.F{rd D M4DJDAY}Vc.S56+h82Mo"Em{v]Y݅l/ŀF_wA'93$Ǎ W&n3>' DLd*͉HkL'жtoTuhm:U$[x?7Žs&I}F^@ʣ4o+x 1\$ǚx^zG+Df㚣?/ >ܿ/}X ?"f/Gg畿A9hgHpP.>^ /`t^N~YWvvA7noǨú?Hh{an_>O{t l?0 ,D^W"cF{HR*yZ-Е_ލ0 Z~|:_N+?/Fzvt:^pp<&e3Y:rԠ>Ď~_$ПwA,碼ѧysz %Rkߌx~诿OcޜL3)'~;}[OquG0~xZgm8>Dž> YhO)'P7wD;u~:oNn6+B)t["k`Xo&ߤY|]Qs/%!T?\pfPq&sժ?}^ңsqV =q\4d FBjZB2ގ[$+F㶐W^7Or}jsؤ\ĘBNg_ ]<~(%`_4$|@S!;پbYISsST]DL3!|8߳%rSNg 04X_7 ,R Az)e%Q6=(TMBŕϺfiwZe/B1wfq̔B-m^5< !sPȥтyxp,AsQH_nƥɁE/Be╟sE% ,Z,2eT4e-}JPv@h`g/FS-ݸobrbv--ƒǪg\莣TH=~ц|gt3kyNwц82x]mfA/_xQ4{iE1O]I>~o&B} vHf"_a_&cv )[(dd \GnW8M?a_O78jҁz,(Dg$$dzFB:B˻׷<ߦyXftnz􂪱A:&i͙ӺzMw.)z'o3Wc_ znۊ/{x.jID<5t_|yKl_5gr'~[H`pQiUed%`eƨ|63ƺTI ڰ;s9Hqh%<Mwgr>oEyEuCyB'f-Ľ\ya?mH>ן.7LF"eK +9`RRD]xH5h;̷HsiMqyr^A}YŽǶ [SU 590XKmZ'*О>jꀪ)ޚdj p<+}̃Ff2$ u^kmtMT~as4=Nke9tDRΈ1{C$^/dCyɒ8IRH:F(GL{o/>tھ[seSF#7!Jc-M6ْDС"B@O@/eU,H}1 V{^t+XwQ'TCBgJ fֺ3 t<`dBJ堲?^=#/rLp ̠4!Ȝ+H/LQyD^5kc׳ ^ DbɦL6G,ͣx.eX"|EVfB ƴ\Jb:tV-㲴$ГFz "I29E+2L}VB[/|'V,+݆]}̼ٛ$i:ՑnqsvTZ6m{~"FɄ=A%SQ=4ulڬЮʝ2#ežDf|Ժ@X{/e'HNY*L=7E1F:ӭEy f})ej;&N3l]fB,)9EDVېb(.}HUE!.T\'9<٨\=7s?b3}F=yoZKfT^T:3l;y!uM7&o 1lƆ($D':,S %$SXNNnR:ΛXWِU~qpgӃq||fՐ!nj{iܱ {1 |l~>=0+uޫn;y888T[ך|[ܺ"Оr{9n#)-m$Ϧ'Gy.JUDR}~P]d6s*-EڠkP"DdU]OJu]6 ani ZgdI+e1,K2|dɱz^sȒZku➀̪aޘt;Jm}S'z`so[tۺEJGzJ;څBd4\ 򐝫PRK<򖜧6 QQQ.[`Ɲ)6gƽ `C\qFKD.b%78Ř։-yΉռΕXgѷ 6k'ݢ{\VW~g,BΦfl8JfdL ++șjr$Rr3^D %H.Ҷ XVkWfꞀ3?q^r}޲])w̷Vn@&Ѓ 3n)* Ʒ Oznllv ɕT,Berntb\c@6 y+T1UyqPa+ 9T0ŸZN*P* @kme`00YY؂s D6TuM_TnX ?5,>q$eܘ&6bMs&|xn=bYaԈ Om WUEtUQZ'CWCwgGnp#՝:CWwCifѕ]-]wR.FDW]uKyX h9NWs+ i3"`%hUj,tUZt [zft%nL hכNW[ztE4*`GCWZ٦Unҕ:V_/Jm/Zr(̴ۼZ[-Xh5L*d?/?ks|44708猪;lW´okM XhvitbKϐZUdBWG 單%]]P ]`h/(單#]Y#+vn< 8,1G*ZrNWv[ztHJ5G LdGCWZ7~zCWCovϾ;tu'{G^7Ҳ SWtetuߡi2# h p 3cVMRn9ҕܬm=< N#@ţãi5-`^_tvtӝ?Q4K(OicV/ﰣf~$!2h6IVm!F-)[tis=cp?/?Zi6& [?as[J{|'h=&ӳepur~;xrӚ[+g[om)jѿq2ӺEV '^ݰJ^h9lsGߺ/q7i?`ˑ0(kGcG{tKng55?EG)o!*Y,~Ud^Jl\[I#-Y1PBB1IQ?ĩmcm_gs%$nRqvug?^%+ђ KzH&g3,T@v] $C lj2L$KNz6A^ں_] >=BRR)JԦ(93DQ() 99Xq> Ӓ0w-l&I+839F(JI8E,[Wbi= aTݹs<tA0CA+&IHXH&! os7@03^OXS+Ɠ`_0K5kwiX!< @#1G$C{?1ef$D XyK!:KkZ Wܰ i|8xeyh"eEk\q$ ә_I.r^1 FÊ,tA ڳ"B#̑2K n;0L&.7BјidXH;S L@laxLq vI*e(:-u5BsH]R2+9&͠f`DSIAV:ծDc${ y,pNUqk ސIX'C$gH!Z:e (Bu0ÔAPf5aJ%p<|(AE 6ZeRAN>'((&'ܕAxcX\lA?ժXTLfrc81@FY0 E&$ YˑAQH>3yJS25I&Hbc3 ` K i尊^f T=O "q2F0[eAH `x%X,xD`&.CC\)L uR#K@d`6D& /Qd%'ܛHSѝA1E#(#EVS04h3)d~ABZO)`!U]Pl ԥUU]%9& ND0 jFdDR + )CRhjn°pfcuD֊bs6VBFՒ1F %Y!$jAQ\6,Viw8`)a;LƃӺ] e.RE|d]&(rhbX Gg15*̐rPÆtdJG@ESBiF66 xTWU[iR)?v1}mJ=6=IYl&l6O=NuH e20S&S`v#2Ep =M6#DL+We2^*S+: EhT$/! ZY.e`j1h6A@F0{ԥv{s 7D>G 6thkug;5T j3`=ѧ@h:'!H;t ā<R(d}iv5$BՌ6!XXV59 HDhR:]\"(tLIF,hu9&Hdy ZHo2"zwlAWCʰZى$>JֈN Ӊ6B+4]?V AkYI$5$YN$egڀJM)Q{mDu,e4Mw+$a:J` ت}AwwAKЮ`ȖZ,6ɵzv3kby1i1b25:L@E ԭGwnJ''f= k63 >7#\;ZY:Jm+5eEZSRu9!9O!K ]r31ie-h:(r]gD*2HeԠʱE =>ok6Q2l]`v5+BqҔ*SyrUpmM w >4F.Lj Re iȣ(ta`Q;p!-PO҅hU.E Eͩ X{umٮ EuШUks3e2 b2 D ۢ]hCmr{:-;V4|-ւ L+[Zy kLV@PxVp* -]Zkċ.!W&EυDq#`5GAsOZ(d -U\ВFn5 E!vކ`Al*\v. b/fDVxЬ'!#i!:m.XAWrh$DipHm,?:COW4z;S`6П`j?.g~-n;/WKWk|SFKtFii0LSBFHN_#d1"jhscԵ҉RQ__6=$jh54PtJ+r>5ZLId|hBo5ƣ%t\^-Rsr!/e3)tty7zҽ|xΊ8}]ޤe:[5ZIIjh e[%F}WV.!9V D۹8⋣7.J;✤Ț,; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@rᐜ@N(9wc\- p9ZNStWl@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; :"ZW q@Z8Vǣwu:A'EN v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b':Ze"^ N8cwJ# t N>ƍWl,'Sjw-f-Ҭw}S̘#ͬV%|;BzF>;w!e_̧3ݡfUgM1]j^W}_`e^&yq.eͯ.S;z~kմۯ\?ypuټ+P{F.)hu]^ٵϮ}vk]g>ٵϮ}vk]g>ٵϮ}vk]g>ٵϮ}vk]g>ٵϮ}vk]g>ٵϮqcw{-5RrZb~7#Cdȸav,؟<-#a:E VDWpv#f0;ZwPnLWCWZ+4wE ]r(tEh]JsW]6}86Xj7>1]7>j7[]*0]}iK'E4+I+` pC+B±ȏmЕ8 ``F7ڠܱ4J{]io nC+x|tu:te<]8+ň0"ǯLW'HWVigQo)z!3!h[ߡ؞|bi`XG!4TjP:VWHW1pQ )BW{tE(7+*>㹈_n`+vxunh| e8qLW_2݀ +kP:wtE(}`:ABůٗ3+CWH0NWR2]"]DW ]\oBWF8ےOWRz++ F[7+P hNWrc ѕ~#YN}7hf՜^k0CGǛY9GE^+%|;Qq곳7KܽL|ݱ|YfXz*4iwOZB.zW& Cͪ/4tؼ,˽ 䂽b̮u]^͚kfjv(>*A ?.#(_x?D8 n_p7n_Oq;v~EWbVm:UUc( T5*|{4˺mWMI6U1YĦmDӛj} {bƺ˺lޕ^]V;楽CǺoc61V&D55g:2s3|0[,cN4Z1)&5JĘ1~ac #J;:ˣ@}x iY G)3@p5ا=v}I֚c+BNvǧcGUNW3]"]\E"n&.GNWc+Byd{2]}ʈWX$R?c΃1ݥ K\g҄6$0K K;]0"q0DVIqtE(S+Oo+O2x87B±ԁ*X+]QW&$1;]LWHW@0"q0Oh'4nGLW]6<`}z }dp7W6>TN(82;ЕdҦN`DW ]\kBWևc+BylO]EDMQhWmgϨYQѪW]YZnb u/.Fjr%7~~ggj]<"emVW檙-wB*8fu>{oo<حor/ŋ?˄['ͤCߣ~|iZ_rN}I"NYy?nL/Vwg`Nm>{pt7V/' 1/)w:?=ܹnk5=.Fo'ͬҷa<';WxGZ:r=>ջ=gOf,n{ZUwO࿴Mo.ycE~@kq6%Y(b2NՃ6mʼnPt&wh[KU <=?ᬿ$~CZp&euxm7MN\g7'ouc<>u4r2z{Yg%Σ"Fܳ ~כ bw~6?^ogOy+NnyUjb5c}s&x0VE[Vݙ̧7鶛51-/x/.5C -.>\ٯ5|W B7_釵tU1뛻z7׋ɛͯԮvC<Zlۋ/>hQ-[\q|0Kӕ,3^2݆Sr>5]$w|Ƃ kGq%ƣ\. ?9 \$7gR,Bbjo{wq߀̢0i =&}𛽴w۠Ѥ z{Ns_[`MiRںK)_b6_\7o_=2}|wEj 𜽏Fgj{8_!%__@>dea ~yH.I_!ii$Q8"qF53USS]Eƻw QstE'_O)|7?Ǽc7ǒv&Kq3eb>)6dTxWr(ɤ>A3L",љ$/;"PŤPgkA{L*Z)0tA"Z@KHga_@(;#s#4!55w~7.]z٠"+^B\b>m?JC;K oW׼c83qH > RZ&kK33!؀RTbj*N)j )ǀ<1@,'FeD| S$UI"l߂oY_ϧeYܬǞ}VϷI,~N'Ryĭκv c*&.ߎ:qB$%;x,Y:G}1vClҞo2nǪ˪u_nwwQrtݿ}g4zkW/~2zƱeg'hl]1,mmh;lcfYyy]C'z=ml_[Qa|zNÄeH'=O{ob̻l4nuU` yDA7fpdVwd|D 2F%K},Fk,BzR XR4_ݮV{nܾ^ouX7?_Nޢȅr/\l Q32 :ʜXCU W#, 8*"\7HW(}8 V6U C;֛8u(Y/ula2~ӎNړ1AZƦL'm_udn:#s9K; qGYc{B5_`L͗*OK޽Kr)Pv=Y]m2'QF(*7DlRK0gJC sla(' u+Yc̺CB`>肳52+2 Ҟ)b ̪nrIW"*doJ(Ȗ.EN;m P;2e) }a!NXٛ@YɾSg?ސt>m9#>F2VIƉ4 tWWpv=\b @DSfF$TKUDl(tAd #ICJ8i:VIs<%ID$:zL>jmfRۘʒQdVz(u2XXR`u'֋ڮ&b%KH&&ajm/T`'P<مP7)FL".^;{Wޕ *ʡ,]hqIykjY|ƎLºAK*ZTRpKsTEƎjvo4?Vh[_Ih.U_d o/JުJAc|T7mƘ7G-b  0M$h7TC]Dw!L W,Jp" sThhASDdm~lQ{+GVW2>uojv~ 6VOLxۤ-[<8r6]t8H =B51Ƃr>;$avhl"ՀAgI+O $}WD*)] 8g˪l&}y󶗺&@B(有U# e"CECG܃E>gQ$A/@SBJZij ԽO1z)t S/;/i7 rwTv|>])t"8@G>`L&N4;T,x@)7MQY6 @7n`d<{+y-;G@(m(hFb2zBժ>'O <)QeOoR3(_K#Hs 1 ]dJ hӞaoصeSI# @S^ه>/n>$'aWgȁb; ){tZERD:o7am-`"6)$Ad#qL9Sy}PͼE K?$YV =7kaY&6nߑɮDY'ܸ&c{FF;\?T(n^y&3h͂ϒ<sI5;8f"HGNIXGY^=}reG-J>yr'FV@:Hi vR}{9`boX7_؛f}|s-WdQiu-"snpz:p:^,p,"R @)lQ"S-,u42QB&t^Zm6T:FUuںd!ZNav%(@%QF%~E1k+1zo|eɛ*l'Xreb6WX3dF5OIc_I$HG!bHLɅ &~ؒMM#гG#q( )IQĤ3J+l%1J-  RK̍ɀX:;Pc@t|J̤86N Iw?%~ԑY1:{ӒE۳_4_I0 jVf*[WTb($4B}0{ӎcIv^K:kȽ')n(V' ُhpُet246Y׀SQ/eHs:}uGKrV̿AcC/}41b4K )%0|B79xLXh凅=CWÇ_S4_3 /NX Jb J|vg_nVgodgsvq}7̆v6_wI9`3w|:1:|r<׃xqw=ˮ\窹ͩhrLuqC2wCG-scqsyB=̪O R<fUZu=̪=̾f mĖ]3z3fY(?S{29Lsݯ ou1:4sѴj<* {l΍V_G?$0^ZUo6q$`f L I)v\(!B"D$y{:`Y>qS3T / %E1 ҵ^ڹ$Ae ~6F+Z|%wn:YЫ+)1EwjnU|*|f?nz6Vk8¹A^A&9Jq%39FZ:o.ȹa-v1,=uܭ ˼@#E*s6˵ޘ)W9c'&Xs䴲3iWNhmL<ݍe?q:9NVCjU`$ ٝW_d-3q*$T@9HCN*0"~źv*UʞT){RJ"N ^c\3eG+!S aj` X<蹜Z@v+Ra@1s#H&v`8%braaμac.U$٤[.^oH/_~.\yEη׏j˧wm}]sK [I1RasFNIN,yϥmeV[5jC9-z2xkZl~Y=7D0Ҿ}vrj2"ׂ=v6iL2Ll@kLy)Ԃޞy Z ZԵt9xV*kܿ# 4#>hߴͰ̲@kA&>H+ʓLlhF,WV ʰTqDrI d !gsRSXrLE #(rHX (e58[l2cS0q(8 /ks/oO VW O/;o<ʋEf " eB%鏄q2 Bdbd^ψ(r5{baĀm.USˍ暊e4)#0m01%}-yHkő6\3 y 9hbTNxO+͈Ѥ 5-!<ki“ZATƂD"s&Lf SFkcB"}FrKck6.#1ggp|P՘i%i"##irǍD6iϼXJZ!P!SzoNI8iMNk 'x6{/c(N~ql$s0ULj<q9fDJBFNP/CRHdBpozitmqy8+O~weHK/6ÑɆC1/?VQ=ˏUpbF sEyu:P,iui7<CcU]=g_0Q 퇿M~X߭?'er,Eg+?"K&U(`cZcP*GF=,>_O.?Mf׷յ﯂4гd1qp&3;Ϗ(Qp)tJYwꋀF¸L-| 7Y>_˳F'~:~aώ_/c^Kä1j2Ȅsqk.6s_<` VPl2-.NCwgWfs" TbZw7va욽=n4zśꟜmW[avUB-LVCc.?9>&0RNWbpSGkyTo<&r (\d t Fer-}9"7EA~Mcff~ߓ{@<+d}xuI.?2):fp*S{O![[-]lx]uu#C*k.S7tV7z㶛C&U`J̇-k|.xVDSPOM=MΟ}yI˾QO+>seC` $a)(=.8__PKd-lt %1 Qyf@Vi2\#Fψߦ4(8ǥ:g,~ #iRjml<,zYg)U{Ptza* "?όSf4U3I8=$ BnrT[.x3RƱOy& >7r_H7}hFwOVRW Ò;f\h]M<w>Hz< q9kc;2a̽0d[ fUTsIĨ38VL[\n&|8w|=#f /02v] moɄy ڂ=tI1?{; VF!ϏBG!G3}VBI@&ۻs5 Uq.UP0/ dQ@ޒdX|l ehPGǕ  .GkV-,e'xg+ 8en5H^o?/V̤|ltu>/_dZ| Y\|˪zUY25 `v UFcar5[A CijeϣO='zj8{27cMޯ" |Yc:GUn2OvCs>֪3SJ\0,]̦hv v-`=v=nc`'#lH!\^Jˁ*1[uzމ0ym۾/>gZZ=OLEpq27N︜ڕ"SWVCx//x~fYORFK,Jý&𝄬8B^ʕdi&9J)Yʌ6.ONɝ%ɝ$%d]YsM] !D}2=Sc"f<ΜdLzܾr/P*\MLLh$5BNX- 28K2؎dQ@#7) "B!7~x%DZƮ]w-Zh)k!dw-kAHTLPʎh*U@Q*-=]]I$WV$X*Ut(`tj#+ +k[tj ]ZL:]TtutwxA*ŢtPwR)7]5\oj-Mt %Cݢ+ҀHOW=VZ+]aBUwq#LFtu8tE4"&h\ ]DwʮzzZ!F"+Q`t@ \hlJjneRVEZpŵF(5bѵ (W(U RHBW->s@DOWHW+}H4CD Pj!ҕb@H4(>@+1:]C+͸T1V"(WGCWV%U=]}5tŶ=;ALK3yV|߮fp՞]jZPha tPcNGCWU@X P*Dz:@"k].\cVU@)uOWHWk)XDt%UˣѮ} =] ]1A%"8 pi4VU@XOWHWHR;sZ0I`:`tHߠBr"i,}hiVw^Z!ҴS]8- pe4>;@Kqʀtut%%8 `P<[ .!U@D*!ҕi]`C. -|Y@xC+0F12WGc Zf(uR): ޷\Njolkf]*z@(" %\-b+@+| %=] ]Px+މ-t(ir%{z$5Q!νeKzM_1eLu¿@Ǣ,ϒ :H&Ye=g UZ ٩ٵ8Z\[-/?LfEsk^ʅͳpa.AS'mWaƎGINfȳ B&adx(V M(e@^Uޫ*CJBzFC\ٿǫE{3\,s_zx^\,o/p1E(aJ D ף| ={mny[W&4 su22gNf7l8- t6?& x]av- kvYp<c3I/ofEjZ1Jopo-B9n}.Y@9Lzb.ZRqJOHy~-鉕ެ2ژ;RǷ"?Қ)lpۦc;f,%J{V>tюXiB)O|o|vg)^im|Cn|k8bd$37Y;S3v*P-م" ".w p0܋^B\'Rưp=fɌ5Fzӌɶj)(^,`q`{ߘ" *Tv蒂6ǝx  ]otyjKq<Վ EX8"f9nI6?.t:cj*ζ*FaX-5G[%{ &=PP NE[Eq ^qb'-^nYVGeK!Ϥ6@ koV:mP%t"PBxH<$C٥H:uGp?bܔ{yesJrXا +J;ACnOW>n&fK^$o*0o>}E".nd#Dh6SHC3KFD_AyLA5/L rXJJd 1Y,-,xGI4 4FIk֥ǧAjäʤZ H|4- (2ASmˣxM ^ͬWs7(/4 &+LN0¹iVvP ݼ{YKk2dڿ ?5_B{#"#JRhL8W**1JבAlmy7NKw] ^ɾd3AKӧG!Hu_aT84m2 [83$\T<'+;=сJ}/ KEs."tԛS`pWKFONh>{4u4Lmv>pG_mun܅#:_ +}DA .ÿ m7@Yç?_#P} ` Ȍ&.S`5fj SjAYϜOF6p>?!hd[7KzZˆFela>3?u\cܐnKI ' L.r|:*Hn"}?ZsX$XǍti`C epTqDU0bQ3)V2""& E #(H8 e5tk\:mLnrlNdzSl)X<6te`P+el~FDzeaB#n P Y6a-FD-QӃDxh$`8/ RHU2>5Y  ~ мKN B3daViCA58x%%z`.u\v@ݔ5۬yB_!"ܰ@2c j,l$ @whL;<dѳV$Ϸ&yG\ BD%3BocYsچiJd7HtVp~as=0r>,2R@RZJt1vZ@с[E`kuٴXómzGEm{v}00#l^jC'.^dӑ{f߸71 /T/=4|>Ag]?a,Z03ݏe8ϊJѤӉLUf"_8f1#uh(Ae ]LX7 9B j2 >~;YR^]/37?0?ns~l̞ ^)N`3wy=+$Kd2Y a0{Ȼj -,s۝_v~X-|;gy*40Igק[Џ d* mAT,>qp݋d6zY>K?6Sώ'ȍ7ge\7yq]#(xtNqE rDY:1ogi}9;ϻ,OM}8tw(r^i×11=?Gp73tywK/򢫟];˧fb1Av_3RVcs;ASv>jwQ;i"?9[Du5V?] F)9 i vo\_KcнZviE-Lj?}l.Mu]u)~LrV'Aú같EᮛJ{Q,J8S師t,,~6ѴSBJ9UiK_G:yx. ?IX9M3c40W(6~Y巺tHvl&ez t߫Q*ds') LA>lnt= FZxҿ)jl(0Ɇ *[=Lu}A=,iA&jw\遹rKyv* >!F$5R,)͊gQBs.ۼQO+zl0(h\AX<詣( K+3^PKd-l OQt  u8UjRV;e+?۱cLRcF Ȳ41F<6AQ0=Hcp[ؒKlfovkF[%gꑮ ڽdE˪qLk YYr,_­VJqRxҡ2 UF+.j2z=`CmF3妟8z_8ѣ WRNQWY0nEϺqr=v-ZdXޭ\noUdn :ZYK0X`"k"QKQE57V"FMsFV;|tZ|)1[80nv` cvlUB.H.\r *MUÓ# vIOn/{Q>*8JI/bVGX^O?GlS"|C8篿[*T˥gsh~IW/Sv/>Q0ϏK&+qכW˼z?g&Y &A`y*H˵_ zL[ 003*.qORv4V=]_gZmZ rԵc-NM*K`h߸>jg;iSΨY)YL`n]y^v~x9.eVBC?A94n+亾Gϕ]X|+J7 UbXL7Wߧ/~~/0QWg?x_`]f'kQ{^}9E[^6u.'| 95p&!'KP;KrٹJwT.k3mbb.9fJV֋̙nȋM i2ۭ%\ C5Daπ*Ly'sE ]g\tgR,}th-< #ED-JzaYqqk^F(+&qBD0 G&xM"821"%2Ohf3!Jee9 ϼ u69 (gJ(띍:`JSΘ aeQG|ie'YsF;s5w(u7jg_Un#.3e¬:~BO3-&^c1S`aqhTfQ`}CבI ?Eڎ(o'Js +*lǠ4UJFf#PH={|"DUŘt",b8nAiE"$LĈӖ?{Wǎ\,^Q ܷ   K#E7Oqfi͌Ѵ![aYs:'86͗ a9UM8Zʍc,u%5cbٜȥ0@uȜH0p9T6 xB kU B41 D,ʧgc9\kvv/'}I9:r E%(AybeB-o;lPp*yq@adj$wVՊvgܑs ̝e$6)zmߧ~:Xe[%(*FO#uct݂vS AnMi砞A=(xR@ւ[2IdXf!P1hL6=]=XM#wgD%U,~AKw_ݬӞƻ{B+rwkaI~S]Ef/<>ŖXv}Aݏ_r _wm[LJw~?HHH+X`HK vy~ZQJxyZ TFQ}[Q@S@UU[bM3P1V’~حTiʫ5ݟ:d#H5a+/V/;u^g#dKe0(- E5IaYil 3; & r&m5%HJEب[뜧Pi(uޛ~Lseݼ/E&4h2AyUNJ{k=l(kK.uYWƳF> cǁQ˄mpz;oE¶ƩK^B%lB [y.%QG{xFlW0WJ茝K/cϹЋ6f:bAPk5\jԑخ4sqEK5Z?: `|aPɉ+Jn$tv,^/?^qq{sm;9dHd`H7@]]^Iы-j8 E)fh!$&ݽ nd-:`ػaϽ=tie*!Ց f-]{؀UlKOK iHG0(SuQ%t{FKT|(S|(S|(.1 <(4Pk FU|OQ^E =9`HC/4(3KWKW:=8#le_?[9R`Ip>G>>;TX2g|hqJ>ȅgMBḳ%ug?\:=1N#0#gsć&(-V..r&Im)iCMԠil LFpl!l~p5`H(i4`8;OYTUK1):b~q;Lfrr I fWt(Cl.8 #n%!T1̜ g9ro4M,eUUZuܜx(\.p*PBTuVI#)F9Ls` BB\cp9?Aփ`dR&4`!,4ZqNs8|F3BUn=Mof7=vrgז^ީ܁{Po%9[٫gٻǹXi$~L%g^]˷'o|S -9sxh!GqæekcM1,%@6j Oػb猅!B:,.ՕkFu P~yJWggwܹZ{+H.Y0⒛azl L`n6ǍWWِ{oFdw׆"{ a">"?ȆzKwFOoz?l\KYa(SihYa/Hu8mN8P};>x̳h2嘳%9  7ׅnvMPZXOـCi2bF.9.b4F75%6Y[!g1M֝'w߈_з;vBY!zONHWO[z;ӘjLiKTRPںLT}5TmZ48vǖҡX쫼5p%!A)✯húļ!l)ӱk!8ₔ'tT/FL!qjC{,E\Ե ݆P^IgS!Z&cbU\v|0^r{_0/A->?x/?q?}u~^|> C.OhvK̀&GV]-05 8b,`̵e +23Pۊ]=\@MZ,Cr6lFxLζ &h9;um D\cqSmOڳ̶md9>v}FO8'JN>y=7%dHZ>Z2IQPl }H6*W?5;l?;,-Q=tDХti]Z@Хt]Z@iJЅ,wi]Z@Хti]Z@p CY7ei:>(e(Qj+xA̍Ck)A5>YlJZ`c<-֡(1._w 5AoJWEdFr949dyZ){<8AW $>Z.C-6wE 6dVG<3[_+Pm9ԟ|`>qY>),x!՞e=d$~2 6c^1̽e|,U|,K>UXqm:}+>G4#4 RZj?A6xPӹh>ߔF 3J6ZNlqtvWoux·Ф{ҟ]>[]e,alM?h' k%N[qC=kVH *pzOb >?| Fޗ4h\A_Ȃv~m~.rkn}x3}h[](M9Ba~a¿ i:u2hrk}N7ˠr}%Pkz=F7Gi_J gEئ 躤I#Cu]@~BnmCd7=o ӿ􈗮Ps9s;ͳ9U} xN(P%hpd/ ą*6Qe#c%F99ŏyJ{Odh>}'_!A>Z+^ߛhqAym6&([ȜSV5GHqFi&Ve9Qm3/m!V;(aϢaFe*:+#R9`S">:Fy us(L*1tN[^2c@21!Fg292@.adS5?g›דgm``IJPV#h<gmEǝaM+5U=ӤمD6Cgʧ$HȔW6k5]MZO o5mg ?ls?Jqc~n\5MZs2n~&J29k_5ӷvNLhEBZ(nUxQN=;YI}Xή$!&b3*pWڤLU˨1f_pTI46'"!%5S3ږrHUjq-mlImjEf 2Qi-xRn!/|@ˠb̲JSyp 4ZQ-^@J)cCd;3YhvuFX%]B0rs-pMSe.]N2ck- <]M::!WaCsZ ;&fG1R'c3iRVsb(|z~{6mXˬ7vAC.ownht{et3z#'5 idC#çwy~/ 00̷ܻ9"lbG&>kN.{p{*DsyˡV lQH~s<ًPuܼ&7rs3ƫsxɽ{ޛ cnj2wDMj=$B{:3EgGtq$``6L,q (\4dYFMXmKF=l7o.2[Kab>2\Yәy.TDD`5qCZ/j-s|N٧&]߽'s /KV<'LObVLSWf;`&[1^}!ʤVH?mp'-׀|2O[-/SFiH']o=giZOov׎J}a4cĶm @ crLa{}°lܔ[6?WpϥH7mdBx Oɩ0w׫{uRrLW O+jr)@$#2R0#!FiR"Gk⎹h p!Z! ϔL1Q2 ےT[㘆lЎGvG(Y90xt;l+ z3-;doPW@<юq ! g %tD]\sѺS*ʻ@o礻x7lZ־G1x>C(v|7J[p IZJS{D|S#䐛,7XBN4NsÙùTxrwܿ8mLkPz RyA :8a3i2``z x6h.Qk)lg) f}LE4NH3dEQV$f# r|6kJ?@遥~ ~?\l|K-='0]쭳[}|ڣAyA!K7 EZ e us,I,Bw3 /`X,l:)'b0ҏ̗蜬vRZr ΑVő6w`])!p}@UnlٽRdqq ZÕh]P>gDC Cm&b )@H"@&]cxFuLNYhSZK{IJ j p)I|ɤRT ΧDL+saPj=3A ^IEfѨ!gM"F,^ld!PcitZ_Q2lUMhCg5CC-I6,y"+F(bзSЎE=܌{ڨ)20iAgj Bt3JqÐ,O@O"z+ [؆78_ҮҐͶm-KSlaO׿{jvI[KBm=_w3߲޻))'2va j:+fI~odJ،${@RZh.ґ^6NBGmׯVFzv~2hp(\YWYj7/ꋴ?o$Ane)i"OW pfި7ڋwtwc~}_N2-\doannڤUj'X^o&X= fSY; 2K5+v6`iZhί ?w[VdQUխ_o˺}+?^2JX-7me|Yeq-s YyJ߾rR:*ׅ.}oZ$*@-~RF]no{ }7q(uNnj?qqs_|&xM;.̔tKXиpG ogPu1KZtXf@8/j2SŌ%B}:EBSxw`O.5\QKyȂU"!Cf#v!|B.hA$Y+zSEĶwU3?^MWP^缋 Ex(,4xY܁@r n BcTF`e%c{W t%v]t䊸m+(B_(X7˄EīO|+(\PZ5YlbN&0&XnK'8=9v*$LdW]KP>Ls P캕hrQafߵ˞֟&úr޵q$ۿ2:~ 0.N:؛  ~ZD$R+QdX[UQ$iM>3A e*N"G'tHrR$NEor|3\w\! ;@Ca(18 UG\u:MVS&*k00`)OT5!Z}qijj}Jp r+(Y|҈" +,z<1kpq5lg^iͦ:Mƭ;=r5νٔs,d!x_//9)%%%3QL"d HAnE S7d/Lbw_-u,!D2V8b,'OQ;pnC`UCݭw_}wy77+띊9I7LL)% /u5 ]$#<Y H^DUv*W&i|8$*,DdKi^lV{EXr~e\JsiZ$ x]&ȸ`eB ã/|~A(u  f.27{,Z Y]ݓ7o.G^om]mg͉0黣EmK{{Z.#?[9γWyԒ䦷%R7t׌oTo3\_66/gV~0u+gtu7*ا[]u޾:.[=py/6 >ՈQyi*%oݪ_*.uqJ{ŇofyWpu]˃yNStݴDYj# h~v]7q->F%'!v]κMAI=oyswooSü,rආVh1~=C5VMҴCJ vn'5iW:?9y\w1ğMsb4&GuUWDpmnoz=sq@]7b N*7unPv ߺ=tGyS=mlhml4u {9' |֏w.guЋYjۡWOFEdmWieԋQ3x,;ErȒdVqgjhL{BD)}dV&kjϿwqaoID2ӘfI̹-$چCqYgC* qQt S#]S?29_v۹y:N13|~WDFKwĪ@C]C*|9 E&A7)BI϶*hw zcz Bp~.мϚr6Gr$w< .+*2Yr^tPq+s5u47VܬG ?]CnXyg&=lBʭ i~r>{M}77z|?'?xغNM ux: m}}Y{`+χ"!vrKmľ߻zOq #h2zIk/O[щG'&&1!2zb}qT8*}YvꎜEg;t t}.jSĭ>%oeR֔ 7< üMPhx3*\/YɒJm_dƮ%vݢF%Ԙ]e̻&S·K7L ֳQ+.$DrI޹(0>(#*,UAZjk ;0[ydu 12zIKLb:9IYxav<{ 6:>>۹ C{%v>M`(c̻뒣 +\VVz` a`( FDW X hNW] #ƴ^o@up ]U:]erOWϐ0ncp#\{J\]XAz#muΔnZ^ںW|ѽ5a4gԄ S |HZb[23&`hhfp+hu( 3i-4"`hu|,ttUQgHWF*c݈ 2 ׎fh-R콫HWVZ6 عjBWy(~9ҕ#̘Av#u$>R@tB]+nOW_];NyHq#X tu/Z{5= ]۱+s2{zsV89i4tO~<^h^:]Yt|J(9xYct%1c+j+ҕT]՝L5 p ]U:]Ie{ztEQ]n4tU*>h4PZ#])嘋=io,fB @u[RY~@Je74i"4Yml >nPqqI@tX\+m¶c)~[&,Z2RPN7^+Cń$KgM«yrZy%VyΜ\ʬJ+D,V:':RR+KEfܞYY&[]rm5gv &jRфt/F{6Œݽ*l@I/GÜznBcڷ`j4+ BhEG:_Y(د,<Õ 1;"x pWckΛxժ ˡGQX)'6+YVǃGknu!p 7u0N j0) AԂ%I%AB\6HNY(g!8ɱD`k>xv1kdcH33 i5/H\ c1S/X0rƋ  3G^Lʰ.)0`^jXÔJ%x"BVT` yz%=@/^ Rt " }E%d H*'0ϊd$AuDk^6ݐ` y#X.-H+ev .'Ё'm q1$E a ko"0Cp$( $^ D q]#j6yЎ2vp d`: (N6G!cUmQ1H=kFV8-5* ΃@w:PDZQ82Gg.63jAqX+F|d)ebzc"Õ,c<9ye>F/~ )dgZ 0][tԖEsi`J fB޼jPR\ ^`VAc-B@̤A@ |^3&-K XҹdRac-Anu>i9Gw4r2]5\Y"k&nj!f=0pS+~)w`)jY:Z4kH͵繨)UFCkn0\LZ~5#|QaA]ျaP^"o* 9py%c~AuP"˥P4LT@=BBЁJ$5@*h5Me"FXe(14qFO*p+ /xA\0 懍9mFِCL*V՗Ko&b] 120+&>K`$1D%N hF LBF t^\ר6"!w*;c70& 0@5a7KBgu_z)zoHISx@Ta c)F o~um޹S.h Wս?;/mm>nql ) )]/A.6SNۊw%^֯rn&V>@zlqqj\N ޖίWplwZM{[P}qwG^r6eu6L'x]g~ ]K oBE uB_(T8c v鬛<%P;䅋JG|J&%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ RU/|@0>#̚@*@:Fg@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)b%E `(``F k:y%4@_ yR@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)^%Y|Vgy~>J +ճQZmN^ +|@_1#%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ RQ^ZF}z+~={Ӭ-5mo//7 o_{Rr,ͯgNj@H7H `.g Rpk.z}~:uh:&K䃫X6Mt͋aqϯyy*og]~_qjo1r/hZs/'z6/˟[aۨ,u>:ӫ_| ;-_}bEo6("ŋ6hv`~7)Vĵѹ?{fZu0Ť1 mkxp‹*>idj ~Rpۇ 5O@5շ-B4{U2Zpn|f: 03x\jfC,891hg`+28}On\%ߟ%MF#ttBogד+^?{ PK&˳oayBo^koO9[γ7oG;PLCɟyЄ_x3}]C6ͮ6ldCDBo/uu)t] xPSIrw)[GVPmyz0#h_s~!3<%/- G~Hs~=}DzήӏJJ7}]_nv.g9>=t1^rfwu`gVY+|z=KW[n&6EN` ;7qPދar=b>+l;}aOg3$^cmMCz>xy[=7I}>x.oFwj6b||=/;|ƿ+Y/rs9!m~v]`owW\Vk·/oظI^W $x L [}0:ͽ. a}T_~,9 htnAx'-l9A1u,:"|hຢTsbcN;XR*֠+g~JiF딲qsu͘cUi coTk_uoCB~Chʁ^m*n;e:o~i{Oֿ@i uU;bŅ.Mr3Mjߨw>,~~Z1zR-o6̭6 ~6кixEYE6J׶,5ˆ=J*0>I#VF337 3>> OvbǙ}m?a;GZw[e$gݛ: ĄO"AFcCx3-? 6Kvq@(#豜G r,;܄*%<;h>7 C]fgw?]?V6nP(oowa̷GARwqY;:(];,uF?SGS)A/4=#t3g#nz\ҰXYBGk:㓫HeeѩL_@JtKdYbc&<=sg+k-]`3%)A*K[`>%_:pՉWrQJE1(>9oFS*?r+}n ϼG0 Tٗh\bQߝ}!&! ?]0]NNڍ~T%K*2SRN4Nא |"8q=9 I; z|TLp]ɕ,+xEvPqp4sh>% D寅'nH /\ %XxinŴ9.e˳޲q5p:՘˚2Z~_(9Nf_$~sp*'Qf(E%f\ӐAZ}?8FgU_ůovϝZuܺ7sF7V{mMf]sk;mc&4i\;+Fml܎8rt89LUFycc9J4,jﶺnnئ/Z_͒ۻ.Lw߶3GlYy;Ca3Zj/ Rbrd{N*;nNVfhd/ddB6ɨvΓ>zLUEyS)(ɺg#2|@1O%j >x}Uz^d)7+E], *F %W+OHH!J~IXĚ\X2BUg@ ax89Þ7b6қMN%"j䈨("RD<$K( /'&=74Rvg !J;|)vF?u[F7nrq_2[%j31T1?+ܥ&-*i4s. :"..:Zh^ظhF"cJ' %wH8D".y 0Ng^%U#ɿ2rƝWq%`9`Og![߯gHmӉ-qY]]U]USb; -1y I@5f>#~Bѫ+$Ua9q ]3E?rJRB7昢gNȒqdJ+FHaqc9h9s)(#csk +"V[.c4>i sl:q7Vu0ZwOA7O\ QʵG ՠOjh_)Ejܨ@p"\^s&U4adaI#|UO\r&_{ƃ}on\FY '|.՝p0{p; t "rAIsݍ0͏>.]5^nhkAxH W %3Q,ΗMCOiNs^XIBsUY Z=!ͺY@YɹN&ibKhQ^,%E.J U0wKq=Z#ֲ.'eNIL$"RUip5tţ ߫Glwۅk_G'o~~&{wߞ~st:9 2 .~_~4 {ܚcj0xjL|zט&yɼ7O=#'kPȞ<`::zQvn.NkwB1t#hFFO3*WOrδZYM^Z $n8:|pY&Z 1efDqsR j(G ]g^4ǜinع]/7q6L@ d[oejo>_&wOU*ro,-#mk8Ѡ ^HA)d )> "7ʝe.Ufk [^b }69 (gJ(띍:`JUXjIEMN#ora8ؖh)Eᔢz'ժ](0LΠ\v  KF*w4*v'lAa;a(Y@8UP1(MLEKS8ڝv53 Zc҉"q9dF;ijI*(ʕCl\Ja1*0pK p*BBX1g{NzU)cs_lqŽ);-)7g `,QWx_]~]EU9̥7j|*)w~wP LPWg^t"0o}Zlj,zdHSZ`'zf)w 3%&{>R@`Ȍ-&T`j zԂeCv=D9jhIaQ)xr ,QA: J;-`h rVCm gfW-GCRB^JQӖ]ލoH6/5!׃nu3\[%pV)/L6L!F0D ݉ %DbDE;C̩SdVH !n}%Hx,TԠkx5n[u(.:zE#EJ&5p,%>bF$`D}+0ia]?rABu۝ass񧹸h!s/3ahIN,_OJsic۝*c:z+tjٰo:g洹m] K2ɌyakvaUkGw'#mnqu[5U~"LO+aK)U4ڂb 63@_T,{G*Dׇp󸾖>^>F;=>-0y>iרPp%bEi+w`c[.Iu"bna|:[,>VjEhMR;L 遆a/UT}>9L4dcxy &>M *j&;Ŝ;W52[Duͮ[Kv(739]]c?O{e8{a(63Vt=;Gknﺧi]`z(` ziۑ< i-l.1=#GP`OtN>F)tCsf[ YHyCsCr*pOT@P L; C?z D,pXY2$Ei%#̦܅cZ|xqpd4jի: /s2U pd \)upDYk\ ^g=~wxc X REΏɱeXf!= SPd HK>sT  (!VfgZT)UgI>=KIA%Q/Rs-̰01?MncѹܞTEh7vrM:U_\R=80NW,Ƴ Pg4hc??n0;=.ZE\tn%MH4޹꾟M\6/W<4] - 6XI 6S!! =nVP%V#e[쩣( ;7|#oS[6Jцb'v 3)0,K G #酗4( 6"e0^@O˭jm<^lr:kx*Ӡ^j C)1U>aWgUu>|[QnOʕ3 nyzNSedP*-b&JaP?+5b b,Fdo~]VYcSl~Ї{ʟ>3%[9~a`ff}o߼jz&fwMбZʄ2a^!X3*1j 3ꄷ因Œ'-;b0`]mT\}E ΄y&+B09vO"8af-j9*IQI)%3ZX?{gɍ2=bŷ b;8#8!q`Z6]ނcl;t=C'BWa!y!\jmo^skݽHkH[iQ=_WQV|jRށBjx!]P]EeXtF]6oK|Sk-l mmlݔ((GqR ZStׁ4'1o&Zָfݿcҧ[{0w˶=CB!?{_on&@oWmz}3z縇:BFbA>mXc7CuŞXh".6*ĺF*Ʀ wMفZ{U4Em y[V>OFʻrz}Gi]3(//n6bzf- W9BJUhKƪ\cter`ʺj65ƺk #خ <.ʖ ebWPu֡kK^ko]5]=q hjs$/Ow4?ѯ=qz˵ {Y6}'tcyEVX @(bBUDLw*]{̽{LXRXUC[ ]吢Nk,ĮTFc: N{=?Q&m:k+ \C&T !]*OkN%`}Ǟd↡_HOl s^pNC?'U3G1G 1!Dwފzci.bBUPYCĢj+Sҵ]\u(Bru:XUb:U׀`>cgrNAOJ]KbSO~K\{f%:ũ8Unѷ 5m[nVڕ;sM5TwU9;&4И'^{|w۾fC;^#u`U z,"?_-קG^-Ͽl/Wf׼]}n_h祽Z4J}]g j=S[B{ mo}gh~91_[2cՖ齏GozR9닼,f{׳ػacZ̾Zu?}>b\*wgZ g_}9[/W}&-GهjJ˫&]Zw HW+Ƶ EWLmbh&+9ו ]Q+u&"+e]MPWv*IWȹK]1dZ)C̺} N0*_%C^>g|3 ښ2ٯ_VuYCW g[.gV s yX'HUiV65MFA5@1b\miK]WLu5A]y ]p9j1`Ej۔72q+ *G L YWԕYod]6Y9WKz>ELo__g=[6M:n LUWtXVmjXc={|=QAnf`/˸A@̔_>RJ0v(9 w hq(mbt[`y ]W@+Lmr]Atn*JWq JbL]WL3!4jB.L2[Q< jC|~yAJ//yJ+n'>w?]_Q=ZwW|:ИX[gh-Bt]ֵ-;J[f9C~z7DS?3~q$<ݘ/_~t?d>i/6Zl1[V;'rt|omÛwB<唌=mx AթA:ǿ]ע-{*\o?,ϾZr^kX+=dg=wl3]^͕y;|Z; mM{vjkjjɆv耇g 89]KE-%VcZR՘҅M0VC}tEhZ/EWL cVg]DWV+뮣KP_>;/Ojv~*yyqvx瘕lyޛy't~u[X!b{|G&6՚׿ ug^|wvF{u7Dm* E+6>/^[t 4)X!ҶQciRyWO:3̳Ds{6wmةSXFWYvMJvK%- ihh-0<0lQ87ki]n2 ʻ^҃ǍRtŴ^+ u"J9G@+QzVQLghcH^W[u2tJ=ssu#] vnԱ~Y|Ci382)]1Cuu٬= `銁+wj(Q.u]1eYWԕNy#HW btŸiMbJ̺LJI qN Z( ('"\PևuE!謫 vֽ%7_[:܂k?#(Hl4b˼=ih2k)}*iu؍z0+ɏ1 *(0F15+Du5A]Ee tEAəoƸ fӢO]WL+x`\F0`gG0Q c7F1-]]Ac\ptE]1+ >u]ePYWSԕ:{+5PR P+ƍb+RS&3PB+(GWAPó~RCueI>Ĭ] =gtG>}2_9ry[j鞆ިB"R ,YP!撕kh!/4.c@IztF)]l6.6Sc]lViCwQѡ"ܻuQH({b4͸rFB6&?-(i1SԴ5n>}q]Q?gf<-f tQN'p]1-uŔ.?gʃ ]0x9"\]I1S`> ,z=n%];0 ZG EWztc* btEQӚҩ J# )"h]1OF4ZYWSԕ+Q99z1X[iY'u5]*ABkbtŸIZH]WL\uE%wvOlz, h^D>ey[|ihqi*yM3ejͲiN0{9` b:DSSz̺Ck%EWFk1b\'ӆ䟈)YWUpPҘ"tŸr6uEF *z]M+ƵbƮ']1eOD_̕wF?=vp{'A#] #;rE2te[ *-HW@}Acq+kAZg0 >rꊁc+;0Zt)f]MPW&DP VY)bڱ ;SFB+6z/EWQhκx*]2Lan"RL1B8z)TSp; r5fPu67vnPS!nH)  #C1@jaBy3Fvj'hh蓎j#ehimn҇<0 HW+Ƶb0mЩ( *( "HWk+5bfrg*Wtt9ϙ Jb򓎙Y|/GWǹRbDZWcj-kMl4+̺zl$銀]13RtEuŔ%hɺ:4>t^(EWD9Rj2t55q]+ESSjȺPQF1Acm=ue7(IW`q HuŔ;&+hEr B9Stw6hG P7j)&Z|fJYԴ7pOՕQF9b\imL]WL鳮`,+q+(vYWU4芁]a8?IWLA+s'0=vp0cOkG  #D12JYW-z@gu+=0RtŴc%hD鲮&++I;+FW]1mԩ(8' 2V֠F)"Z *u]1YWZjOJX'FW+'"ZTBueu :Cҹ+ex.?1Y⚷O1Πt┒`Pmde/re4XCj/@&2 րk ,fIiIW1Q8Gj*@{t 9oXjݱ=-U/^ 3[1)wŪ}b{uE]:iigESꢤ*wM#Vش"(R]CWuEm4>:_勺UmOs?]~l]—t}xߛ׿ɏ7~˓O=!~ZV-yO/ON^]uO[=c0Wvun|ŧ.1Y>Vm7}ϷGWϺ1DB mdіu6=BWjM2z!o9{le|-ɹx5_ϻ?ߎ>lnӦ 8~ڞ'/|x}Ѷм ׅ>ݚJX$̍4 0]|IGH?t6ZA'b \c8ލޜao]fBˏ'5J[tTjYWq6&*:_./GF)JqJO1"fWo-JEb8珻lA1whD`Wz2Y>rIarT[,i|K %[r8p2lw#Z ZYժӖU_P LpRdXQIɤ,T1yW~h 7#? ^[9gνpg{0xfP"%sȲGFl9 \E|3HҁEfAG Ch-SD$ׁb ;;#8 l2wl(Cgq&ɵ֯4˫MկWXL1TeEEv,GERA6H"8ީ0P+cIYҔ2D.-KB( rPa rr,Ɏe3rJ $?¹"E IZQeMHy*b2ڕALH Mj|CAVmJr6_sUhb^"#9imG,`b[ώcY%0>8uTEfl 2'SQ3mB03%1$ a:8sd̑2=b|$=$dz"Mxy>k׊FNlL]F̞[4v9:r5kEuq~151`jJ0y</7eqs{69ǝCDOR&%+5T,vL'ͯik=GvڶM}37ήӿ g{0a 'p[*$13/źgA҂IѾ[vZx?ϪR#gr,_LЇ+> sU>5Tq:^仨lGIXUeL .*A*HVeHyϹޝs(a{qN{$fqP-" ` =*,|Yg( tjB 5$< rw;t>vFP>[.xRT fDŽ%OMzo'Д3y @rmdqtM]D~VP~Jy22*dL1Ʈ"縠y ~㩯/?>P>[}O'7l(*č}qY[Ar۷Ĝj5}-QeG_Vzw1[Y{e{)&nZ2:1lTJ!1+G!K*0hpp!(m0T3jkrJ9V1ԄNc.ǤBOrנPgRR]kȹÑΰJg3θ.B1GՅJߖW'yY㇢qöLfד_[d GstG̑r"($2N%pTِIS% uaE4 3MM2h wd_c̥CtY4݈9lvL̾hθc"apd\V9ibdH1Kf dz!@rTf?>D,dBœQ!]E͑HGRVE2x߱>쌜6˱\7E#vՈc(8hă3(S Q$bԆIbIJ4zIvی~}I$6 <=RVF F,i. 'Xt}nW׈HSH/.pW/q|KEVB6iA,Ӂ'm2) 5CdJA/>^>;*lsCKȣ/b7$qw&xCLя ܏͗0=IJGm*,sJ1 btys:WOaGaaBx4Z,Dɒ+IЂ2\P .|rfs:̱O姅u#UVWfyEɣT=Z]W]j}g'D`JTuK)}zH28ͺB*YQɵOj>^Z;ԼS0?gvCn#_7>VauEzSj7Sm}GS?W$?/v!ܼ~jyzrw}]JNڍ?0<=_Aףߛ?h:x/W}y" N$$.ښH m3p9J8C#XFhtq=ЋmPlmí;jk_fVŹoVJÊ石y,U:Qҝ fuo|;5Ǎp zՇxѤ[G~Lz;.j04l~/FqUEMկR$Yj&Nwy=sWo{ߕ_x|wo^W߿ =w%rr+ $ C叟Ѵijo4)4jЮ]vѓ+[zT{:sĶf_n$k驒!uevr:Y/G:{)b)nCs9Ƿޤ ` !^/(x!|9  ZK0^v0fF0'p k(0W*)QXR"c1 Q_FDދ"P+6yuJ+\֓j=dNX#>gYי Y"ѩ0ئEne%*7dJiF20|Ty7!xY${F97݆`-Dg(%:L99+ CSORY4*:(W9H=g<8FX8d)Fe%,sl(g~oMhȽBL"FD+rFVA iIZHf}1cz"^˾St }ۧlm{p=3#<>eb1){V'tì@eVJ~B.*'|r3%< :*g+5Y AZ|Ή͘!el7jH[o[XpK/GBBBuk+J<$`D){ul,nWqZpNPX )Jqrz`1| <`8>-,y^8,0 xKcT4pcD%6LͺED׍q2kLB"82r4~Dڧ1ˤ0(aoIƻҨxN)cY;3:O-٧:C"C )[-{5:@SK|ČHIhƈeHG`:?/E@ w{a3s͜w9.ww~5̏`b^{rKu|V^e=+*l7SsE ۶@4}9 ǣηXz3clh8ܠrxNv@b'a\|ACϽ8~=?^2a9N#IM\Cu>u0'X}ޗЙGi(=dט&=Ꮓ~KL.6?I(*W4ۗILo|z<хLL;~>EjK;chGnﺪŴDuo[mU$Kvk7Hf}7oTi!vytb8Tޫ e/ev&ӡU|Gg']49x̚#G d']gz}\$[ObOä!޽)wJ迆QY"ňgJ2ѹs%|᷎ EXx"7ҜY Yf1q4,^<4 CUB쯋޿n1s-SK RGcK?:oDOpp:/KN[,P(\S͵, &H6,"a0Dpz?SޖrDž1s6Z|Q+2Aq|%/-#՝ɔe5b>*[Ԩq9wK$U֌-uO^?=~/ѝ;e (B 6:25ݧ'ɧ`l*ɒ*Oo?pU[/ 4NJd!gi̠οr-)fU%+*mt'jH53̔TcE[9Dq -]R%6LߋEG)= 3xaY?Gϗ~v(=R|=dE\ZܰHw|m~# +0% ж*̖T*%chM =g\Dž1kl &uD9hkcvL.|j3ߘʲ6kHӫaF\M`[cR;ĬW逰,>9HysQ b+e;FT.֝hw;in#mJp]~6ribw?毆F.PW˾\l!'>@(}_7yz=_3e=^yfZۖimiq&ӂ ;„eS塼L}+nqc`bҹVpb#V6b]V W=ehI\(wޤ$ U 4&՞s:FY^1қoѷ\uĽ<5=>Gs?.3 m=aoޖ)իZJrH|O|Spa Uz;U`Wtwpweןf"yw?`x<ظ0?%axEoϓlX yު'a L n68bԜ8S;:%{Wq#8P"8hŸ/ Nbxpn γe˵ gfٛ8waHPsu⚕w3GNn0٬Ddf|1J Ï҃bX8dg`/wSTOx3b85\f w?JjRq̕4[Rנz)e(R %_Y8vr~0HYe3b'FmSfԸh]QǴ`1V%1ZlBq etۨQ1,%!oo}t87(r^O:FaQ>awTf C'p/eQكιQ:타 cAf^¨QƸ9Jpwfia:I+v~.I=Q-XJ"Xs*+*AZԮUpJ!9t* qc e1p;gd[+ S +J(ImǮ$%GWW]/jm;&fۅxKIIn\d Wv=B JfP]pը)p"'okj5) i Z "!WI\%W -Fx*IQ Wo(!Ux[+KPc*IKU \1) $0k2)pU;ϮpIy1 {XTt{62}o?1Ʀ!yPeB̀W٠Eb]s]7:> {>s!\0s,{slqm빠0xMf9pS%>xz.=;M;flJtrFd'nWKxv1tpbm8U|R ]"]E.i_ܝbV/;zb! ]"]ErH0up] ]E=]{v'AWh?wpwݝ2y}Kܫ}kVgH|a|&Nfϟf>GGy"C8{?[&qyQO wי)O34~~%l(Xѵe)ڗDުu^WY%*|8~ ,Wk$WW| j}~7p){3J%=|K()S|q:+gIdRtg}l%$Sr!U-х:WkUnzeq!Lm 76M#s}atl5|gܙEK)jDs Dzo%[Z`$MԵA3Z Ct%'+>'ѢkMSn޾} I5Kc8W[1ևإ: kg 끒rWj9в#b!;Lf 0kD34p'-Ռ iTUiԒϯDF0S"^?mɺM(Km 7 UY2dClt)CQiԎƬr61b5)xٻ0F>8$]^❤Ӝ`AV#>***-"w@iLb oDXqvL&^G;=5(!Ȯd܁E7BnTzC{- ȸCLAA&󭋠@zNG,JLhW4 ~DUuASPbt,,x:M;*q b LN QCF:\CAΛ2MC{X +TtgqJAQ"Ł.0͑&A(gIwB"=dH_(4u `UW $u9kh ~vD3GAyǒk0 !ѿT(`5R!pPBY3' Y,K"1A! db9WTU/<-dPgF\z ߿ܡ;łTM%c6(!9Yh>3(RAUvNZ'$̿2K/Svۙxa]/_o՚oks^oj[U0o3?8;5͈ G 1=*4*KPtd㻒9D%@ۈقj2AVa1 %O|.rB/,J肸 kɐ$RQd"5B5 C04X yyt/1 JWLd:[nmG⭐C@8U]UBNu~d}:MOU;:@vTVB]+!Hb!ҧa~x6`RݼȓB,UǑ+Ob fb= eDA!۠mAJ<.J3nWHnLCQޡlN@rFEK֬kdž@__—b{L{}췜2Ɏ7Eh x@^,4zV6 `_D(vV8Y:Z:5WZsLڌQg5rFCo bL]dvz4\}YeF٤#AI'? 9k9lzBg 378w(uR"TP=`C(uAAJ@25di3 A)顾֡7q1옍}u`?n+AZ964NB0Cʟ&o Q0 bp\jsYTQc$T:9Q\J#xL 7A lU?<66TƀumlbEJ@:Xf=Եj@רMg&=Lb2Վ *WZ@׶䠧{[bѫxdPi`EHN֔70hlCiN3 _B\ ?RčtXzGIzӠlCi#8 8%/C)F @;"@.* lnT *vKcb`XKEv,*f-$R&T$b4uIJs't\5wOW,"$R  R DŽR/^ܭ-vp7WA:QFӈE,/-5?|,FFk~:U{:w~|sX/bn4H ^K/o6V& X~=~xu;Bo5V\oq7ׯ ^VyNk~Wm$V:["B;}mܵrڭaHLF* o>#wU$난@ 7uѾOC)ȸF'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N q}N _hr@q@#@^O'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N quyTvIN 8۫w?.GOjzsoڄ)|z Ѩ 5̫;Fw櫶Jv|2=߮:?6dÖSJ|W}wmK_!a!#}v~y#NvDѧ5E*<(1HH$jtU]KY$d"iY"Nw{Uo ~cx7`P dJNRzY)Y̘C:j۴!#Vuʒp*^wJ[}Zf]뫥(ڌ-2m|1&zG'l9=hNv\,22BYp(A:1E2$;HuN)]\SbbkJQ7 -` U\ژlxZx92hčH7g&%)t2\wJ(y; ~T*ͱ\٘Vמ@HWp$g?wU]%1tО+0tJsMX`Qc*%)tL(yKWuޫ1@JIq+tU|}Ձhy#:%+lKWt=0C1tbJh);](UKW/抱֘73q ڧ*HW EDW g+YS*UtP|tő&I D6\FBW 3PJK+I봚?d(VPYB m]s i ͉%L5ZIN ji%ҴLUhc*危*ZY{2YQKWCW +DW l]%ϺIth)}.$uJo]` *S vFCWxϮPϞ|t `~=gZqʚ%a|]ᖮzLҬAtWʦ%tP +BԴAtWUB{(Cj&>\@g>O@JםhE&9 ЍWUBU PbZzt@!zo</=sa[>j`倖Ӓ_sh9LNu>0=Sx֝Z۝>o0)O܍݆׋z꫄*Jy%9 AC`=BR&<*Hl$0dRVO@Wym^ʉvAb: Oizl&F41v(A\RoⲸ,:?_ Vbj) # I-WpˇMDo&N=Jm9PIݲf7> ˟fgZ/ ] n \|uu@&}Wf}~X@ɰp43*54%Xg NA &uA IzɪՅ>+_ ̷:3]tፙܷ`p7_f&˧\{`ٍK*wP)~Kֲww)*Z:مNEɴAiE"$LĈӖXJIR`b1*0ԁ/^*m@SSVE#^ywۨD RC$,R•E3) vP`BĈ7 5 (2Gs Wl)e㻲*3)MaFoe`K(qf<+#IE S jc謉i$%Q_%'ӟfgI@4mi_fQ$ԇ;e.^d4,g70gs1X+*Z:Tis{_gijkhjf:fPDl;#fqeH ))h?"> Ų-ЬT*jYTz y?j%)J=Mh*&`ǻXnR O6߅i <$iF~㬑Kw 2)9 QC`ԥyJC Tw._<I`,UI=~qlBY}(fϳڌ9#+>8t^nݎV0"g4`)$TƂd#"S|$ʿ;1ˍ*Os1s"`D c:i 'Ox@ƂǨ2$Pˑ 0 ԀcPs1 qTyb"{&CD)RRpDpCLDK9 Q >=͘!eli҆ۖ63|gK4fKo'`ɻ:}Hm-(yWubkTa8$c1PEU{v}8@( + Bd>_Oݤii-m`}gZFwӤ _`^?>E&q(0wCРi`1z~wf Ǭ"3ҌY Yfbw@ !so/COWp:g|us^ pe( \&upDYk%ՈyټFuކa:Jd`[D[ܫz6߿ׂVEJ+|;u>k混UTz Rv*ɖ*)\[o {e+Փ.K{ͅX$/[ڦ8пL͵OyU-5⠬[d^43{\^VC`rwUhmtI l4w0> ?Q>91NmԽk\a |;%f*99D4\͐eٰ?<1;^ C;E;͝Rkb#uy']͸;gC.b7-/h^q^&>R)ƴ.RU$|`v4̫8Xv ԔB D6sl짜xS\xVI9 6XI 1 ;d1D@ jXGOmG$>cgJ.#:8x*u52帴ZGx,:ak[W.VmXϳ\{rq9=q=VBVR& 1A}0yTQ͍QmQ'(_ofs%1{80v` cvmMmœɅє> b)!a@`3Xxti\8C!xqMGE9R*I$V+")fbcd-B K"fk̫腟!d 9љ-ݲ NZ1 `v=E8jQ`X̐b0"Q3`d $Yl^Å f05iR`/D^G*#R"%cR(d 3âDc  ( G,ь w=YٿOw=g[[%hwgZgx??/mn4vjQ;#B_."kco #N` %"}g),6š5"ϣzd+ ƍUFM/&w,9;wtu(\qqxa"Q-rQϷMRToxU#shk>=Wg8Iӫ_o~}ᇟ|xW޿qo]f#<_+,/; v郯W?ãUGF ϣ[~JV<K8"BY]lcK{WaG&u ]Q,_~ "Vޞ6u4iNjwC,J7| {,Ja"|h-J~oj֟92e ֻ>l?2rg 28%D~pU3zEj/źMl%B3JH-&K"0SIA笙9CD%iӽf&Uճ.:%D; DJMR) %*Z1xT %DD";9(6&+]":⡻Qt>ހ:^:Azg8(%_jx懪wez p>) #-"@夃*1ɜ}۩b nzyZ<˼18ۨ`"r$l.k <aF qNDH<$OjP}:*xw,$n¿Fyq^Y fx]^"?bMAwc'؇^}6?p,_=qlͻl5e;<׈W+D.8bH/1flv۲'=V1b4)Zj6¤3qxA@=*!MPk sYwlG|I|I!_+Xcl1AMTS4&F6艣/k>X`IvDNn[tWPBAO|$+ecI7!)E3xk9J* .e0'0%{L,<(UژEd2\.Lj͘ ֹ`%SBC"URs+a`^?0{N/E(֩쐏z *Ai[5c'TۊbG-Ea\OJ4PD/$Zq($K8\(=+/)s1$ICNHJc28?Ș4tblߔ@V;X̖ Gݗ/}ؗa(_^#>F,^mVުZ] (VHnV$.pcK+*"p-zL}¾J(nP']C@$Ky}$I&S*ihITP#,h%H>o6~m߹_[;ND?.k߶k+'ZX$Fzm@B CiCB1:BctLYpe8BYrIrW߂u#ZVL'T}iRo]5iȕql//U;#x0+aYW`KV&@3tkgȮutG{,OvZ$XjT8 j&I95-dc&g%HeQ!e0HxfNemVY48JhcD!r%Y@7FK`g;Kӫhf~3q)aF\DOm4R)#UrHQdkF͞v ) GBdRSP*M2,ie/qRH.*fj3*gH9 .J?I?I?IJnB@z'ީh2>E"2:i1EboPL_J_ȼmdˆ&dX{ρCC*+?:oL+q)~DJ1p#x%# =SNJsݹdE)⚒ʂ1Fr*"'0'$[fZdr,a"euNfrM `'(w[ )K$ЉM ۧa1qص# f8]jCDzÀc[H(e4ˢA:JA 4)' Q3 4e(NJ%7)'Ӛ_O8'j8Y,*nQTɮ%HM2DdOlQ1h$F3 Ic(2\5.$: yT!Do:dyEf|EJaTid,&=bqbXXld< ya,= w]gf&)JsΉ=`e[;|:'d$,F&#@Q@U $8)E*8(`\B *(ax Yٳ,&W+A֜*1QJ>&)QJƣbv{{x1N9'qWww[/Yȭ? 9Ex'@ޏo?M2%q BT+] Ile*!2=i9v&cJ6jYX1B /$0)jH&:pR9Id>K9vɟvEveqkf2v8EEUW6MV/9?>e>@ Us;G-.٬d ެ++rn߻o.x[D-/]y#C^JbD='B *CtǦ 㜰ަ'"x4MwȦ;aSa멉Z*LiCih60Z.77>59'{KGXu`a ,B HFbP<ދ/%=277%{gWs=u}Vu˧7!&ҾΦ0ϰZ$$kEhz4IL^F6B}+TEkpr4uWcQt]R^~54;"B59k`c,-],%g=\}7ps" ՇgW$0cpm>^F6ixˣշIiH}\d#,GW(.Pr,pݷ;\e) HvDpq4p\ei:\e)% 1)\ƠRpeX*K; f)pdJ(xuDpu<,W4],%W=\=A\s/y_,{ro7-:xQ;m4ox>m3XĀSO  ~ճdZ&NW %J#$n]%\՚uf@z*]CWlæg1e-dՃ+ CDWPmb5utmc$qzj ]%-AtPhlm+*%-tUB;:D0[DW `Wƺ5L(ah]t]'m ]\Jy[*tP* SHOUh[ 2NW evvt&% X@Hf_#3 J壷GÏSQ/G=x}3m&G"]C_}$$t Ok.ZBLQԬ~ivջ[dld/iv G'G+8:i<d "B QEGK=*eGɧ<ʢN~=iRD%욗`)]:8;71E}+:G@UFeK=MrQYn7{ۭζVu7/;f=b?׸E=Sa bm1DZBIHgY+ejLޭzn*yW2mnB TwM;/D1Եr[ղx)̮kWHٗհKS^V%94|X Z5NnM9:Y4".0Sw6o K{x2#Dh6WHCsKFDoA"ߤkIA8P?NnX"UB~I-\y'e.w?x5f#uI ^^po2U>,]p!o(zUElR;?ga5UEmor.޽n!ɍwr!h,x)C^1%QǠ$!.LSYC]}fun&!d.Kvrh&g@d =!"}2" ")qJ7Dd%0lnY*ᤶOR͒ 'M}{SPѤTg7ho\V|:a~P}Jg6.fU<=xmg 9'ax(ō2‹4-.ؕUij.֑ݟ=Z̖}ȶ]/po!p %4FW  T6d Yzd_I\D/{ >$:N/&@`cពNnh6}$ &@cj^#[~TbA**LK0wEIӬV-:oL)A*',%o`ͽ~#>!~h:=T\w`OrDN0!lzJhUN%E]A.B /i(0Jmb(;>~Ȍ&.Wj ԂniŴȆżp6|Swކ-&6͵#DD_#]D?=fÂ~p< ݲ ԴBg" >|J`3;> '%":nNSjE夠X(#%; X`T9F$㑅H`,Gꥦ a$EV 1Ob Óq9 AA3o71`x1gFLpLO]J/I5'3|g7)Q{-DTCGB8Ap7V;(%q1“0BJ;baĀ]*FA !CeڳۇBT5@^\%+^䓑_8[%pV)/L>L!F0D 3-$ek9Uʢnj /V,Y+C3gkc@pţ'g9WAz?N 4ğF{wIt>s{?N_Oq?4mћ0O9%Ǣ4c-feΐrdH2bRe? B !Yo?+dg^>GbZb@xnper~Ifsg3C>1#hU95/.(J_W&!_?16hr*TzгdN4K!\MWTgtĘR縉 gO r^|AT9KɆdLt%+ŸYjJĐm R=ދ`AY$hI&jRH\ݓ|2ʰQ[wv;Q4U@bnvgydzdw? PуclA NF@IQfS܄ng'S//hG o;8)LE1q"2ղf/7ܓ1&)2Zr \ˤCp21" Ҡ(8\#iu돻!.0]<ֺu'w7&Ɠ˛p:7:(ء]x83p׃fv&"Q]+ptkȸ(1'uGbH I$;kѡm n'm[l`}veDJ٭mb܂zn[vGnS*z*׺GlmXgy7j۞3q>vnz쵲2aDD&pͣjnDh:vF]j&i:7\~"fҘq[:f/.qXܮ\ޗ('7`5X΃*r]:Az){$}VԺ|G z_sWmƷө5vWP H"XJ%j%YD1l 6FB(mRX1S^c󎐰ra&y$TViǬQFF7 V=^#v<38 6mb+[ KqGewC |yIwbz8sb,n"7ħWi`Yd?\42ᨵF9b1CSIDx`91@fIa(̗tn0K&x:R") Fa'T rK@GA xxuB@qOð巙wgD5a  AS$-wJtnjp,"68DޡhyNªN۟kh*EKC@X#_ˎݏG0.|AE.~J)R+R[=Ç$,2 $S3s^# ƴ4 !.yI D-IZ IO4@At\FT`[YaÄdn_JڃAm8C#s+!:] 鼘>քsw}{dyj\/oF8Gx'EÃ\\"lENߝ$M@ySO{cOnt[7b{7vi1i:k=.te>ǧŮSWvmnWl=ѲFDDfmC⨣Y++fUFl֓mVi4~7e|i;I_w߽~yG\أwDe[EE gUе5mƚw|W^ғKk7..cPwr]pΎhO84]t1t+ĬSZeY&`Hםks uL GR͇ׯm,WB-o {I8R7GYVᤴ~Z\8'}Z6]9] EmeLk>4%o[eؼ٢+޶_-*Ɋڊu囘mg3kٱB"eVq4%kf Ԉ1M!rİZvG3Cϣ$4fQĹ:G|Ƥs>ZKGOĊ*;96&qyWFt]w}Sutq۹V :㬃J2y?:m'*x6 K,mcU "H+n;OmGI?ێVJRT%iX2rks(@T~{vtpj &2xбa`Yh#"yW60!צ&y%_G4<\TJQznA`h( F,̂A"КGseTJZCiͮL|D_731m8Ccyf0=aN3-QO eܫkǗqoX@bD#l q"XB0A4$ Q}4A.)r(rU2Afà12-Bnwp/ :+&GAF&4 ѨPwsw(B=ǣ= o\3kvWq. ]Mb3,~6|>S~-6ܓ_oy %pZ|$f2UY^J)=q\iE9!5RJ)]M@r em 7[3{K)y笂W[w&)ྞBDR@<0"#ϲ ]Lb}Mfe O'i:ZueD+BLht[-H߽o{ݻtiŖ]dώz7oK'r5[K{o[ \x_/z$%D7Ev>ʢ:ԜM.*(lG}\u*Ǜ!76^ylP$;gXOpչ\"Ц UN-ӄOchNd- +I9Xz=U-IWsXb|"f9\9x|l-V=L^+MMSc#uМ~:(ع96o4>y$Fw_\D6Ɲd!Cbhɕ,\r Uq<3nq:̡Xgwr88n6.^ZbuwJ,gỶq!FQV ]S)y1vL&/V@#Td/ ti£U Q2=*^uA.Ew󖁻>w57!+8*󨃰D`NavHav /hy%(y#y.2=tv0]g4?D|~qW|?aqҮ,Iiޒ4/Z4$YeS%g}E5xQ^6:Z@iF\M3tyU.K{m ״ a6zcw0<C"0N%ܠ`ɺkP`U2}? Fޥwv8:/4A gAɰ(O(ƌ2eO8O1nF{OxdqC Gk9Lf5tak~]e -ӽQytZ5nەѽ『 +oQLZ`ȃDތNF@'c3 fD3lI*OOQ}ш]5"Tֈ0hA#: lRAFBj  %@"9 f7N $Ǭ рE胅F&1i&'PJTֈĹEDړ^7'[gQrWXغT^bA8P %QK3E&@p IQpRXb*HhA/އ^>6:x*l_tڮ{Ob)Wem("G~Hgm>:U!"4k,DdraS<91{ȮGU!dOBH"Ȝ(I0_P  -2D^i˓UΎͱcy_h凅m.=M/W7]o)|>E9sn_rxꆮn\tWnt5Idw?ͶA+W7~/3oߢKWَVgwmCLe{.ΦQs9Kw]žեߗjjsC<= LB[6pG8ܼ^Ǥrt9k=F*+@>57IM{t[~O$]<Qt "@?`1M06Q9fx a&YcQܢ0r[a>r:s/}*vbP V-qn }"w\ߥ~/ Vc] }!ayj`:l L `6W fWa~&˕&򗯶ob_XCR mB1,YBtS>l 7co O(ᄨpvs'qo %<$ gLeJ] dWwf:gZ+E JCy嘒LB{fA+ x;sed\R=BLH2t{S2nNΖ|s;Ժ[dl1ṗd1F$DIt,0g^ 3P _R&Vi㹳%D!RaB#`Hhzbcp8z[KK3^YQ<2b&V 0 ]t(6 w)W=V J;:YbOU>T^'L".Vdh 6:#5v|]]}ocm1%9 &"ܘ`d or4P& B4;boc-iye[zѶ~Dۥr`P|p 7("se^rLPQ'$i-/2%e7?T|ۃ1/vi*x6^X8y?}?`вoV 퇒׎s$ \HᥝuhqI{xV/X5"c|ݏڹ, 1={oW8O'R9'7 t2`;CWn;-VUFidOWJDv2\]ɶUFUOW/^}~gq<3]=fKWB <]=%vѕ|]ɞ !BTUtW*5t(Q=]BRH%BUˠ+t v(k^]1+ :CW.3ѲUFEOW"J ;CW.]UF{g5ҕ`Fb+bL](w?Oj5fU :2ZP> XcIϟ߽L'ր9Y># FV `O]?r+jHh:ݡ WȮtFxi:Խ|4-PtL1dt2ZzUQ^#])evI]etJ+th)UFYtzJk`KtL1d33;eUҕBtjZthyUF/xAthǮXzfz\c?ǡ-+R=]=1]+ ;t*hM QR^!]Q# XU{%2ʶŮz.tŌ5/]ev hj;]em[w+n4aCt+pMgbNW%%=]BxNpT.'efv{_r\!VT)"domړd6p2͟FX~Y%b>=%w*l22 /Z8Γ',YO"VG9ݯZ=;`'q^| lN*\6\HkaS=1Π ErZP7Զ޵9ՕOefqzWyF8hY,B8N-*eZxLQOFsQqygLD5'Q(U|Td\`1BRc8 6J.Yi1z髶ye<כ?K?*Z'_]R.fY./{o@5М(sYeq<0|v|tT,{.d1 d3f.Pr€WAq=ipvUNCX8\؋%,KB߶y~ U ~Ci fJ)KV]Yիnw fe#zI /̈́V()58zLS:xh)jnHъkfdVKc3{i7yc;`ȻeեF^=ԓն36%ά?EK-Y/L|2=\=:_4^P27:Ђs% m) ^)éۢ>N'.n)>#=]5'\Ycsay8&\鰘 /M j˽砥Ao/^Ų>I~So^Òǯj3f Kw>V8aHHg>#4FŐך!ctTZpdCir EtQ#H@55C1RdT KF(MM{߈תkr,sLÓv\IL ZԦ]:m;]꿝R:[zߵIU?wՇ[xyنa˵B7b`ZtX gaBH671iTfLtfLt+V$5}1<13B"ڀUR$GJ6uU6F|y|Co3y369k5Iba:wRvP/-ou906x".v>,$X|کZZMAňQ }P 54jeJ - 3. |DQ)EoPsƓ]b]R7i^yoi疨mҾK?FyWڽuv^nl۱0C|2E%Zo4n8;y*Ww3{rlյww%\(+S%u"qiTԂp(T1jeޡseT_J (ROkB,*KnT963ʅa.d=~S.\#e/2." wg;妒_Phr9/Y$"1R.SZe0 htƪČqZ õ΀w7ŨeA#X. -C-)z146ʇ ŏ%GaGϢB xI#E0D l[s?Nj4c[TF 3"g;{LC7p)"PhH k;%.F\mxԼ3ri 7bpdh宿3JIIWz|zX]h7f,iߚ q+7 wzzw޺r?t6D{vZN2l̼ro.=2>п[ r̫÷_犖ŋxa-1G!J交 3L *UX`w[gP@8g~2- ӂщ5#kQɔRqT*-'RRAWk@=2Gx@h:WM%Ywb.9( !&Qz39'qym}~Z+JkNtAn^^N LƟ̝X,x;U oi &v6\ofX[fl+9ֲls<;<}7闲n]@2lLbRy(5ucu&'U9c\߿ϦcEK )C Y.YBpC(cl&>w6**I;@zHת`^t} ц]ԥjkk̊`Yd*hl9+e#ֺݗgV="2XrZu L-l~GYfno() FL skaXJUѩni=hm9Ϋ|.Pxd()܆^s6s,.'Nk3Hl!^s(͜G,#xF1Rs,,/QP|CkVSnkP뇪Bl[nKL!M܍2)ؤ=C3Lw|r Rԩr踬|cr^1vLcW>n]yFj9s6vH)V]+#cWα"k?2ShK+L-hO櫉eNpۼnkֳuxYvj_/֛ajd0e{ brʟ%w(zV//? 1VU:Hg6z?['#n/)Gn.TO7x2[-3t>xԤ='@֞tOvf׷8'9I+nzjמEgt#j󋓊qqref8\A'揹pv3g9َ(Y-ǝKgk"]M>G-{NBL{kw_7lZ{k7N~T[k-GNk\IF׸]P9&"I'EiF9])2y]z>);*0ݠф:kUz)C`e"'![׃>>gd,^߯zNkv~Nk8Z4-{#?RMRDfc}褴VUөdM6IQ .R.5zp- wO=—PeZT1^Û1;66{cW\Qc/9%C ^>{ o:@CVLl$_;'VXćTPHv9!LKRGj,(t_x@Z PD$9:d^S0ܨVٸ0ⴱ.2@^8!Fl4_6E-Y@H#@܎mCG "Y ԏoT?/lZbܑjm82)6Yw*<}X pC$*J]Q43wtKg۩R"T2==ѳf %hv%)%ر-OG<]r}:{ 2y^I"+%Y-":4qtN&,0zR6TM>; ifMEQ^yHY=`4vMfcf0糄 zV1DŽ :(Jjn 2Pq::t3%fdiB3r䆍RC1zBI/P>C[W<׮6b];f᚟$r )I^^^ѭ `\]ʲBaТau92c`t PuNc-и `UʦT~6P36xk FnVn $c͚ bӪ $_X&k&- ddej)s f)7iy~9(FaLMFzհ;[* X6B ZxX@VEF̀|be2. Lt$KD=+4h5[A=-#FX6\'A5󫆷 .b&?Ϧ|8]QL/U@@dr)(HqC(;)jI-$e+yj\;Y];WD={czlyۋ/.nNTГ 4F $g$e#`12* \bbS4ir@@EEG {lU[s"5+ԗ1 ĝ|"Z\Pդ܂LfSez&>UsϨk09|_xtF״2ߏgE'idU&9y+(%|BG:)a緣N sYuJXѫSִN\=MϑoXlXG쩒un</4#0(XNN7]@`ߡGғ|ױ0۳ G "]y jj>^U'\g0:NF DXPgD*9՟NN6ys݇3S]_?|>QPl]iuR뤮9n ~+L_tצ~>zЭCmi:JT0ҋϢ(, ?Ϣ(, ?Ϣ(, ?Ϣ(, ?Ϣ(, ?Ϣ(, ?Ϣ(, ?Ϣ(, ?Ϣ(, ?Ϣ(, ?Ϣ(, ?Ϣ(, ?{Ev3o ;̵͜a%":3찒rG<&o Hvz9N@vs0É(`+m#I9vKܘF{ iq["5$eUHI,JJ?XyEFD1oyV1HĺR,dK/BQEI  lk[pMrb\:*2PSb+},d}dB֛8wBBc!lچ_H^l}9z  "՛`?ZaYyyyyyyyyyyyyyyyyyyyyyyycb)ik kr/XsV]ǚF3ĚSޱ׏o$3(s]P6Z&0^Vr5e q< pu}@DmT$=̟JnBLn7:B{9bes\̼5Zd0͎HaS2&bT>>{D&݈tdK%2j q*Э(]4F;Z6"g[s9Ӊ1el% S+ QrNYe+dJL$KIxRTwNYwBfB}wi^L˹%LTad0K*o(Sc ̩e?|2ۿִBD#߹Q a^(r2!4IKL&ɲz^H烡P*m7${R+`jQ >J V0 <崀Z0H0p>ތ\U\ªNA1Fj&S"vTj:wltSnwe?\NҘקn8:K×8՝5 jXw5-3jZT6Mf@b/ >;zsp>k\4VF:odWm,jH:iأ>ߌ OF-B٢lwS,Ư0?_ٱga `s$ nJ}4Q "~-UhK}\),ׯk_W_=̜듷߽=i J:E ZU57&U>[= ꕬ&w{sYfL#@5;H'.-7/kCvYjjLF1昵J 5fB{~tIqC̅[2H GGW߭Qy]0Kԁ9iLY QNa*buOon 3a>20J\ SŨ 06w4rQyk~{yFqI.3 J!jIhv$YX,9"ݩ955IRY:utaeaI ~Ҳe.\6X9p8%INP6}R$J(5Y(2>mR RxSfZNrz9əoj7ۜ] Ools-ks4!9ʼ?rzGڧxJh^Fh:ecNFfUaG~vdPE#vŽ\ؑ ;raG.ȅ#vŽ\ؑ ;raG.ȅ#vŽ\ؑ ;raG.ȅ#vŽ\ؑ ;raG.ȅ#vŽ\ؑ ;raG.ȅ#vŽ\ؑ ;raG.ȅّ{Ŏk .ѹ:;2Hγ#vȎ 7zȓ1AO<ɞ 2TQ0H412]ɲXx'YIJg$x~ N ߬I 4,y/8WV\毭4gym hpmo$C,s!U8@90XAz,ePGbө-$.8KϜV,^zۥYkp_~厠@Q(Nf88k<ӫ# kaz-L0Z^ kaz-L0Z^ kaz-L0Z^ kaz-L0Z^ kaz-L0Z^ kaz-L0Z^;̝͎ 1w an?v⤅7T !r>*!bZW!ŃFĴ 29FbaX/5glX:ZOc`+KkAY U4Ԧ]<.'M|JXIq0=.&8:] A'%1L>HļbqN)8U$,k/YiazgCűw#j6Rt;*8 {/~?y;x&W +Ko> R(|xOB10IJ ˟(x<x& yp.ϦqQ E1$"ىBSp*ͅV#e uXfyRU[m. m]D70H`pGoF4LŸFſ|;iD~|v2oOFVдwvv6M+ߎ0X!FZ0KXΈ6e}wITzGqA^**"ƫO*ڦ.Re8U_]quztY+O߀xԴa  A0<#PbBU#'M^dd))(O=f5Ibv4mVelZ!쪖yIP\WbLz&ĝϕ*ʫg6%Zr\M&Hd2;*͒ܦxQvQ*Ue ttq$,7W]Xذb"YG;8}deێ"We/lx֔}xR&z.\S/_k6٘avyG 0;vk/_كeܴr1bT/Y5fIܮԐUbw%4lNeW`1օYܺwˇa_ч^sbnmc.VrP9H`;-tEy;W+򯛤@3Kj&oFy+5RW\x9JF"ީ\u2at Ga.'h.}wprGy'xa V{Y$0j?[wOd; wfwV"fOib&M蜟Ok*f6E{Y }FJgL-pƳ?Q,"s='>9 I%$R8ˑ0F Tb8';s|9G=< O˺۷{[`8XɈ2̣' \/t I"q9iڑ8R+LF"Yc?Jg`I^^ٔ6 XtY+G)hA[ mrp5>zY{CW/Ll/!{/w=R~\\]C9ڰed0|<7~?#z5feA[\qa ^ۡ +پۡjPJ 3dSH]썺qPZ gfʈ=RW Vto+ g[~Ƀ]>5 M Um[Nl;` ׊#ɏꗟ_w5*Bofߑ55G܂{@X"޾Z#5qQ\kEM]W(TEM?C5{P\77RV)%eE]=Cu %쑺\⚽QW(H)Iq2/#uk? v_Jч+ТL%gl\ig'8-a# L?}]fIl=DUPh 1ln:`Nik&OÀd_|;-;4_ qb\]4Cx:x-ȸsv.I,V*t{nl&22$; a^nCԻlpLJԘ[{#R-s9pL'jFJGr )?~+Y#}nNuUL2,K\̼5HԁuJ)m*UBvwHC,0;ywL670"iFI߯-ْ-ile˞nͪbU*;I;pnI"'AMY |V <6 E{F5V۰?H [ZZi- ^H3r%47 W!Fd9,=? O[<ƵGA'2 ZT*gEn+c9 %_cAVBN%dOZ $+Pe LX>'Dh}k<vxW(q யFC5I_W Ҥ2џMͅW7U[!_!LȢxu?Fy889vyFQ>:y}G[a}u36`v&tiyb4\M&ͪF >GO=sp6m{6bmuf`լR˜ R1Ϸ Gƹ*5w/ ^ǵؘ`fX) nW``JkРG|<́dPDaQ+2IC4Tz*)d[3}`sYxJDSO_k `E<Q "+AЦ9=O?~&L<ԎWe]kÙ[dkKn_;e{%v/fZzpsH2h3DdY%2m3*]];%8V k=rn=n>{ZSmSd(qbF.F΢#gأ0"d)YJ o';w6Iv9;&ȸ-BVo7bZ+=ط}8p ҿ;+f˛]Ȏ d |X;k@`Iv_;ѷy/Z /IS~R{6p4 U@E E ]eÙܸ JVv;̏LΝB7ʒf*/c8-)lqwd\n-qv{_{SiBLJs{p 1rIkMόAϴgAhX3{y&C-{Pxz^1l]D ,Y 68X%0MM9!uHz ^w6Dj%֏c0xj=aBOVXtZ{D#3bXaHqNcBIƈ$h 03Ncz6^<7̓ik,K)IJŝ7cBJ:E͙BJ }/Wsro)9ԓ gg5ްY2} H,Z2Js`Y/JAC7ºjZl]Qw2ҍ߁Ȏ.xM7 t왫6&CbiiFbRTڝj-,DIǫvPԯuN+kkn>&;q|2@?6TX}Ug60[|uHk GNTfQŤ߆^Jʆ?py"q7PNy!Ampx2;&4BkN+@ʤ8[HӋtR"v4k .tgV4,Fjȥy[ ofضԏ"Y'4"ח#mFzFMІ!l7 /[ y4/64YI HyP^/oșuO/#,vR:0p/$}h0WƆ c˶@8Um!G3m Gך׬doA ~9ح_ U VWYc,%*6+&W9c AX7sָ0WE7|q-71qΕ`^$.vndQ̕<ޔWI,&'; 0/]?h p%IS_0<3%Ct$̝$-͝٥??:6~f{ 8 ϵz}~I"Ve y,`tOG,yӖMܴ8CiD`|f;NPA!H5\xۼܾ\pۏesgbBD±$s$83Z`z%J:S$?=#r::x y P<9%2#?7(L{͔`>gsi=F))ڐ'O?F|RTxcG".Tݡ]xAtk9J ލxR_kWk-ռbgC| (>Gə]:+jy\'D55Kj=T[+ Xz Tn=J׋RuLI`?[YAG^xS.vSMKٰ^g%'W3xj&nw]q"ۜ|bBEȘA'eJRJл$Ov5fWТ H. pr_3m;h⬓4?ꢳ FYȬRZy059bo{3ݛ˲պHZW kz Vx{T$=ϲh2LIM hiׅu/]i.!Uqb4W#D|_T%7Z{=Vi۵en40B.{鶞foe}si]nnHqZ,-xV⣺mP;ʨ4'99)=WVȠ:@6XO摞yI\B1!Jē UL$9sN+'{[[&=3k3b=JLzDs}l}QG7Y+że;|o=GjӨ<KΤþ<:<lhM.Z)kj3#Z Jx_&*2a t4-ݣy-BEd3eEh9X$&!U#@o-k۵"_V30bv:GvM1P1:L.\pѠI"yrQ*o< Hp1,lP J09BsOh(E@ZrHEhn+zNg*vQȳ LZqj,Hɰ>4#fȣHƀvqf9Pyͬ8dϤRcH.3̫Rk^(Ȳf dRJ8v 6_˿;#2izEC?pW`XM>=𖉭#؛G?G̹dڱ>Qh M>n2s:H'ZWuHA %ΛrQlqW-݉ PAY &3D- kƌ<3\C KRO[|y&Փ:91$Hk'KjN)(#S@;3$]&M;kbZ]Ë)Yܪ{v*_wXaw}$U8v_aϡϔRAck#0IB==h.tCLw{4XI:Q=X~DF262O6p dsؔ|JFgJ^#b '!W*qAȇ*=`,YAK7)}e "J]*DϕfQ83%ޯ떖nk42vXa0՛$-=?{FWvmPp\f|Xf*>܇\~-ޥѩ2@z-҃8>,DAգΣT|dEkY{\"bA\fۇe Ȳ.M{\NV!d}MZYFؤ[2\wE[˸ڨXByCuyLY6du~T=\ȔcV:W`gkTg퟽u)n|Egz/wW|upωw_75Z^l}.ͪgϲ̺Ξk~oI8}Y^zgӲ.[son+F؞oѴVav~8ݾs_9zm'D=޷ s}=^?F^0֨! gZ}ō7wKC^\c8L(O/O2^>\rrSSv M!b`Vs{BnۋB+FuW)% tZ!?AX5,1(e|=z|nEYdgYތ2*9!昄3Qg&+.M>m <~W\nZ<,Ի[$aTE \[ a!=;aA+b D}l}1(`N㩑a.{!zYde.-q,Y;F (1ngiIm;^(]GiI+Mʈ;BqO-+n`-DmǨ+lb3gRZbt;]֨{Y)h/-Hȁ> d4eAS2F)K앷b޴&s( O(od*&Tjș¥Ӛ撳F-VF4dig%qvyqʐ5j=%qYOdY#|#R" 57gslX8x) ,ԝ'u>eV]|Yup}}ۗmzcbQX*EKa'WˏuD覄\L"KOo.DF8=} 셬Qi&@g;M5*SEn݈'v[.f-Ԟmutjw F쨽ΡL2ECΛ*茂+bHw3٠UfHOHL" ,C䗤%"QEFQBy,9ņ03gK~mX g[/%"LDDD|vśdmI6XqbK# ;1J-?%Xi4o+nEwo\Rg jdbVy ,ݜiNO?>H\\lzrmff.bbs+>' NWFbQJd.y$ xExHx7:\| .VG-<|aN+ leMlP |Kُ~B8 S#^fΩ\VΩZSӂ, ϩ95pE[r\Zk:Yqu>Ba %\\ٌujԾql=??\]kE_DyI~ZyKqq{Sv/n˦":_ZG`L~3~:j`(h`3.EPgQ_y0ڗ78-?|Rj~v_G@+V/7`{r> a5~PZ?ןt4`(u x3ǁcN 7}ݙYt3ɇTRF c X1՞jgpJU`՛JrlC"^@3bʶ+VUZqu +l W,WVpjWuq;4h60/]# Z/ߌK:|;-^e'\3mj%:mɴeN-W-jڊxTctnYw)mokТ̅PZ^y]G&j vDBI. l%jH(=pӃ8«nW,}fR^{Wx\ʭͰu}K7$X XVpEjUn<:\9/m){fpjz\-Uv\#T 6NZ E XqE*G~\^|ǮB&G=u5MemG\N+h2IAlW,VpEjԵUjqu!\`mE XUUvFMpHߒ3Hjm[v\jv\ P8-9,ةfpr}3+RquzY56hRK) Bj3qIW aj'frQiVkMf;Ӗ& юUrԺSu<ڎ3ĕٖpEj'cU#zJn`D3bXW-VjWBԧ?|4M;u ir>-&5DRU+=WS/3%\`km3b޴+RNu{c;WJK `4\\o[n2v\:WZբ!\`wfAV v\JWg+nW$HlW,WA+bP}U:C\-;A:ALvC;LWʹΙ60l hTatmel4i4}l4i IA-m``qYwؤ֩]lV >KPH[o?)).勤uJ0M4`jlӬFiVYۑ7=lC"?:MlƪdP}4ݪqtzHRKY0 3w¸- p!Fid[j[ÅUY>\ƊpEۖ׎+ka^Վ+R:C\Y'ZH0J XnWֈqE*3ĕsFNXogcjS;X%v\#ۼxX!q,?_/aLjJ! U艉?{F!vq6dvvNIX,iY~Òlm˙FYdW,HLтt^f`SqB`nEƿF/dz%.ӮIӻ ]ߑ9{f~Xzg;3OwZ)<hęO9Kn%Ȥ F&=E_BV^>ɹ-\'u!漴{R:|Fxl6:R^WBWu|zO>yڗ+~S~<$PyE7% r!kǽxcT ߏR>j=\ktQ~R~S#uM7oj1g4,\xg#bn&l)ow_Q`뛏8~g=M{Wyo>:>T[ג,Fv|{#h?.ZkRa@k˺ՍS^⸺T'"=({zutTH g*A sIbv&&>nO-O^Q>`o EH5'My9Ϭ<;?sGAS@e.q؛s%UTL'ɵw1 coïprCo7~۬zg5[^.̚Pc}Kn{'dO-.4OtcR`~+nE@غ=b$9=^aWV#W-=P^ ۽`1|/qNy 5/MU_n|Oin??R>j=RdC?X(Lg`PkdUf;iS]e-eߌ⥔Ug$t>g\FX% VB?$>$UtG+xWsˣ̻DMh0#}ODHL(Z, =M R #H"wH A#}eqn'j<}6-..lѠ +OǓRͮQZSmwS31_6aHac γh DL.I.}!5eQF1jT?s$CBe !(\FϭBG#=U̷l߂w',Owkͽ7k_8ɒ 7 soS[.{Y<"h< BzkxBk$ &Y,1K-14:JSsڇ [.nӰ 56cuq Qa+^UesLTd'G@g>d ɵ,F6ZpIA֌xf)Ϲb%( r,;|b~AcȤUs KĭqIy]ՊQMc1 36%9ff `S&EHi@͘ct) c;;"`,NJwmky% mkB[eɠn:+x#%V3>N\m^7㤵t㤄#;^?"29u&ϖtJbh+r]XQ\wR*U('9Vr<%i,T'BY& +`^YhM)#r#11Ѵ:lL84后-l9FkZ;qj;TSLE{9C&JҲwp9N9 -g_(Mƃ6@$ɋ* ,U`sut ]a4z$-4;K7-="] 1_ Ci^ RI| xVmZ N*5Vrk"J:D* *gIt)ʾ,}Y&bLZK"Jj,q 1BF!uJlUBUv ~SB4t`Ncmx5%ٞ0m5>sY:&^ibi AW>>]S}TiFZ@?*gL*+Ŕ@ YiÕS~ib(b,fDRtd1_lDrXafXbҩ⠸ )+\deNAzӖak<0^}L[/G2U8v?ӡ )Rd::91`SUd; 4%V2/`ҁb<ڍPB9 pC!" Ik^~KEU #D*sDNZbqmPˤFK ?AfOl5"z5vME%w/~4y0Mysr>LtrV⅍]o1T!oo5gpCS l񖘈ܢL9+8mєRWkGe%]ME Y/\KT6dܐstqN9Sz1l;@ՠi^c̳F&gZ5Ƞy$T M ,qL*9n'n(d'dB6Ѩ`T;TY(@&G,ZjD[a4LXvky ,s&yRx4#DmM&1GlM?e:y{5dBQpJA0N27RL+r![nb# qFVC6dτShp+HLZ(KU!ک5q{īsO/Nq\gkZr_4-E/ާ) 4>$B]2YfHf .9 Nn4C*:\vI|Аz:Ëz;&0:Ap]+E?ouy 8xR2Ji+ V s\80gM=Gx,9f`Ve%Jc'hUւ6G]N!,$(t)vas :%״N5NN L=-z:$X(eI +!+8x(ק!^.|$/fh{nI/F IbqS)T:ZwW }!gJXҪ-/8{6% _6#}0ze'16 8~^b}J\SCe'%q(J=bI3=UUu\Ig͇hy5J=__?tUv)f`gP@wI%EJ7nTRAq# ] N0I4K$X$(:id-F!%3SZ{5ܴ+0g62g ۲ң¡sQ1࢑|:qZ0CSIDx`"8` K>wdZ7K1HwD@Tp d 3âpT)`P$wj: KU Q8E*7tu6)KG "iaSJSnjp,"68Dޡhy+g-)*>4E̅"ֲYLaP"!L]JeBvK@H@"aU 8 lQj&U4adpߦgc.ƃ~o"\g>H8d?K՛/{]$HN@ ]84E/&i|Le5+I*& B #Q,Gb|HMÐao0o$3a14[beir5JQglqZŨd\ guُ40f8rkQa̒>ٰ,nNe~~|QqUsZyp&uOl"!q3?$5.\q.J~ *SVS.Ho^U͋__ًWo0Qgsz2 .EoA {Y[C{=m ͆FкY Ÿ9qoҟ&zO7,ɻt&b\irPl%=vE1VrT1w+ń qR+yB76OzlZbY`Ţ֯K>θ 8;',IilSM)m%)@ m}˰]UbVtgqt_f"`mG84ki4(ű6"%zAYiFmbYyI=F:O^cyytgUOelFfUx2|3,f? ֹpI:7ԨIXű@ ?u>`;o=TIqq4z%/C  sFw47*ew^Oː2 /P`79K 08PɈ(:$4t ܴ2<%*2&!27mCFZkCI+Seᇢt1y'Q츕Bc6GGAcHc FJQ\ o+[KoD{xֳ)*M0 t̠MZ)Tp6ȷ~H gDsJ) w A1؜arRcέ8Dúv :T,7<urMĮ6fڬcTB`֠}b'c) 븑N8MU Vl#b,*Hhˆ^FK<)VRSFDa$EV U,Ζk#k\*OtDjE} v:|m?z38G}7 k{MJ?G2AL9? a-FD-Qӎ##$\Ѹc cxƀ]*F@ ʬ29>B|Rv[e0$M=b>n[ѻM(ʜ> ]:+4uO˭۬ǫr_tcԫXRJUPW^կ:Rb RhtU)۵K%Q|E? euCuP͔IB2L-?Z2 Ԓ,yVz`{*K^Ȫdud., oH8]IAZvk9o |7h ЃV5p0J lqJUk_TӢ#t0oX1aV h/ںr F5׆.=а@[ߢ3h#޺>Ф.߁eYƝv Kjx CgWծRŲ<|,^˭4lM :C6C 4zP:6rNF˔SGQNq:H+fϟK.s<$(Fk@NiŗA&FĵTKܟudn%0*(;f&Zk#3QN-ʅGϫ8cS|CQ*36݊k)Y1_NziM%"q!WZ<8M 0h|HfFCp2oB#w=)e7Bdx_B:p( CT .rvӾ`y3qex'w!.GE6ke-e` 0GM,GX5u[mzN1-F7+00i@۩0hzB%Hrf{:%D=u Z~%+Y%czB K&cgW \F Jڟ]%(%Ȯ8bH>% T=ve+<*AIhǮDvE]bWIr۳)~ %?4ٰUH` ̮gjr-c1s^ĘF417p<}Ik>" Ƙcdt5׫[x@gi`0L R}-DgQYj2LiEJ]}W<y,![nu '1beP{aػ >x,P[9nR+R3Eyrd3ѻTswdv ;tp7Y{ɱ6O9>ג) j k&[y(ͶsG= -kֹcS(`>W;ݬ ]m+׆;Q{z nD[cR;ĬW逰 ߜ ;^sXe~u~sEX:p:`.}v}l`~ I~ \rAuW{x_ |O֣ݾAc0`?wג|owZ:hWEn㸊p} eq$/HZY^v4r pW!qx|~U%<cXac B$%[|fLT6?+i6b=yA)=VQIek O2qV{tHbEw2f;2m 2:gI&X{,[ `cke[i8;ٻ6$W6jG^O/Y4݆GEX"$նOde)S%d YD׃OC:fNTՋZ#Ս7ӓ?"!ie&%PKC#{j6EF~AOr*֭,fƐRU6ޕp"IgfhF IblT?jŝ])( ]OL M˻TqU` }4̚\V:؀$XB*1m I%) )7BX joĈ H"/sR[%A虃{#g?ߊ>^ґi}y>@:Eh+F4KDc/gW a5?,Oud8j8P1ʏ2Dd-_u:L:pV(PR)FS)=1N@1-XML*㽯%:9*}썜ؓ%tsMB'c\;/zgw:ǀ>1bZ%W >dJ{6*WlTZc L6k7zh՜5NcvQ^jN?~"&̪rƠHQ5ʙj2F=T9]9l}YUӕї}Y%,:02 L@JB %\rbϫ, j -T3FowmӃ峏zmgխ\JfnW\V)g2^jtkh!Z6(mPO;TXG1_zrNΗ uu_1mi[oMmOdjS_)CڗyڱB#t=u&g3246,ⵏuuਢ>KW [EUrp.]5 F@L)]O/NdGaQs2#gd` 0 +Y`1M "VmoܡZ.=]Wg[hxe=׻^; 8GZ-wl {Ō@ n {\: fīOWJWj('jϛ0VU MALoaC'"͌XA $Ȱb +DqFUD+I/s<% )F[*zL:ʒRl&ij*Y%#d3F݈_CS)RjRAfhY#.Y/<k"5aXi7q,$V>_M.R]uFݱ8_33KHB82z̗R3)ޕ Jʡ,\h2а<ڍz# L`]Z *`-JRe/d$?*䢦c J4j!de  X5h)d"9p9GhL$P&۫U ?19wHOxOtm~M潥 /j-zHqDGOg:c?F}{_%O[佇{Kgc i)g/(!A*mgٖM8BFq/M m#}s! [KE#reermerfe2FDH2T}EĔTVӵHxq:ɯr/M)4$+Ӂ%\xy'aZ>~vigOdX@ QUJ0³ ϥ*[VڡWeTjlNH)ovzz~—_kB^op6MЖAYӥPV4EfhЪA%US)!rsP sw ڀQS1 kar'bV߸G5%쩪͡pTcyk#MWo-H%*Tr੘OC6=:AlSv"D5ipX[Da &b{#wL"$NZU M΁O}5E~FZ4q: m}'vns>>棄<{QW^ӳ 5][fŗs΢m8=jx!7T5r;}}XzoAl}ɘc$Q đXF8wR`:*mmKmZ8(!YPr\Ȫx~+ B %co!e7қ,mg* UϲPAeEum;j/%G'E-,& ''O'/[Q $ZjUŠF$fJY ߎ>T :O*CUBbJVKVG] -*R#i%Q|Ӿ/|CڽQjYÈ7}E7dPA[r0gdm6(%^aaRdHYCfdX-95 YIid2?1!RM.ݻ7rVq(qD=KD=JQ"0 HRPqt rqbR[_` A RU `ӬJ`N,Ud ȧĚԎS$c껴So었_udl65$[goPh{f\ܵsAW*UUep YPhuF(B.E;" ~vQrE,ϸ_:ϾCDJi)8O_ΰYdkxdBzJX0ǗG st`0ǂ42*T,IY$ ޑHVUՇds= ߏ 䬒Ee1? J{.GzAJ+%0#꺫h"55ɻ>gS>tjM68_.^-i{@UX80U=>,B#xINJ2A9bz|>;_'1%lKt:M.X\؇A)meO~k!{JPMX_qqXs?2d]0 jdҭMd%c%6 BL3p+5F6:OO)O+٤TP|umպ nۥX&L\I|q UD;dG uǡ/IڲvEk+ 7g5~n)R5h׻ݻ9[&i4ΆZ>ȗ7|xyܹ{?,m.R};y]c{^tW:W?ƪwؚookۛЌd>Q3]?* ~1I0FLU&BK#l2R%b J4E $x @Z_ jV >АE%uɺv$ԼXx2##H8@m9$5Sxچ3HŢVL^aâH'R(9oW#cv'}dA\?98j[?W~j/9_o>x{}vim4#&tVNWu~d6|j+=-Ut̤Ig:FvOM4^NVLy;s_NlrtkqFw*aEvڹVp9WhӥNlx}HuhQ'2VW1*/` 邹az?Ytn乍?mt૩0kjo;5 L}z0}.cޛɍDXǃptV?tkGi8;nzB4kŘ+2bkZhUOVmRx k]V8:R+^ݦҲk(<,nf`˕1"unw߹}NߤQ4bx_{.|ńr/׵Z6Os^f!Y[X0 JBieFYTɊ7r#)tݨ=a{d5-9D3 F&Sd! !|%ko})$K!=jMЄ,6E^<9KZ %A=q` x7>sc5LX5@-Ac0MwOu'J۱@ɱ8g4\qGs{|ӎ'mg#Y^Dڎ(o'JsNʀ08T2b<&(؟4\ MÉW%*Ƥ!Dfq rHkv(!g%P!Bl6[CNX B84!!GAq' t.c\q/B̝t3df QACbe$u48 C 2Ņ:kqIx׈dhr(ȡI8gHḇ3]`n?9Xbh`nG9^z4/@x.Dh6WHC. ͻ+SBeoIA[~)1_aH#+050E1(9u03ҨM%T8 y X{))S"!&"Ŝ(aH6Mcy^ DAXXY1-3sNQt^U>@n`18LCK]kpgjl fleC'E0+Ug~*ϧ~٬E{&ףx89fo޿v^\Kjz'~K:XhNEMD?m ?sOO Ӈ/q&bJ{67es%(3L\͹<Q]eGd6o_wPi}OβAD6ߦsb~`:';kU4 X"pఓ`7 Z`ĢrRP,Q6;C rG"uԔ15h#2O놝Y1`da"򵘔ȳ_$^ݳE>0a<^M=u?fQߍg񀥭KRW{G2A#=&גz<+~Q/q&?v_naû=aY3$En9#̦̏ٯebk/gaz}N?2W}$9k| RGcxJa0_ף~5Oz3LY+5g9Uo{ޟﱜrje+ 'ׂc&ߢqн__ӔW$IJ^c/@̞2m"c>q.-uɺ1X][0(Rj$+\5O˕׫wEgT]bi/ $JU[^URbR SiI}> m뚶@4SRȓX&zRfPMo4S Ӻf)<^#xBV/Չ$$3ާ~&od12c漖A~nv=vX+vWIχqzпݼL1ag;r h/xuwp5T.ܱHU)eIMT|2V-Lq9Lb0_W,=>mbld OB4w*Uxnu[/`Jjp6)` Q(鴂9y,2&tF/E_6ǘh)!mr#Ĉ6J`|r)R cѭ5zixTƆDjĚ RۈA i4 w`޾֎vaO 5Nx[ A~zI&Syd+-b&JaPQɼ̌Cp+i+3[l]Jٝ-X~lA=|z+HkʟsO9]iذqu[L\^+k)L@yM>jb we=NpW4Hw=nICK`I`Z2*[WI`ݞ%W!JU*#Ep@\[Đ*Iɻ%W#--1 &=+Rκzp%4C1<-x ΄N%2qp(Q6] 0}Op?OzL_dۗ~,MpCy O?~_{٭jLyt VwoOVylhp ަ ,Q"u~fP"o+rzߎ`o~-sam[31pxsSsqⱃ(x,D=ض6_g%֝ݓ[KFާ<) |a&k^e~A}EjsR죒Ȍϒ>qc<χ1f jxztC/_?D/+CCOK=hh`8-5; ૜Ik &V0#)KF\DVgw1-!&Z2}hc4kG9w6x6lJa^iGjd~sVk/;梂ٗ[,} *6I b *14ؖ~_œ=plNLj9+Q"։"jֺ?➁*s=+5g})~3jLbU6oڭ0wCܝGŷqgU-D?JUc/N'7K$'ʓ1)re wNJ+3j?!wU7lo$fCL#ea15^bK7zd&b( ֠C{%w6;޳8Mvﭸa`~f9\VCdzrV^R]c_?]M>pV2C\s#<9OFkGt@Sui"G6NcJՐXhm _PFGR*:dy\Y@G$pߞk~ťj||<=y]K`wlI~rג#ۓkj.W{Oq0?0=_2D}; AZ\3U49F0] vu v̴ #= BhH0pvp$uh%rR dXp6ȹQTXDcQJ%t)$J΍ / 3 w%S.Ni}joaXK#ޓG,5qfFU$0K.̚ɨJq=؞L`i4|g0+κC#{S 0Pj.Hw~ a%d*źIϻt?| ]-$hvTvD(y(vޗlu^^R5\]/Luvkp{vr56l|X:{tATެ8cbj:5aVmӮQ|7p;FoonW<1C*as5xKrh)J\Jk]GűM/6nntgoodcЛsG&.ql3>Yjn|{{vv}[ݾyvȻy.126ukW]TsuSwaZ_-RsŕU XUtޡF"mi=֒O/oru9)4]ICGL}2Pf sLL`FCn=/RzA_\^e뇷\bWi#Ha,=`a]sۼ1^}&,+H+wJy z(9gwހt s&倱]0 i<Ч;'zl3yL lTʳ,")Y tWUi{&'!J@-ų H`FmvB\GlDJ1YlllFn8Ug?i[~K1 m`|pFThn*.b5bRES Ks "& M4iJ!=߇.3A P~d(Z2D`s2-(*Q>2^GkC:!TӖUoo_2k * ,Y<;)sqs#N(2 )\MQ*֧^Jia(EV(qFk_hm bf1l5In̍d%=pOlK4> z<=_zO>Nr̿]\յeۺ EUy.Ь>ɪikSd!61D@kS~P(TAqAqdAq]keh((Tj^"|!tÀ{h kUoϷ}/tJAGlY3(%>}Ý?J|8Ë}H0ˆ#א [}E[%iR(+"hP5^Ih#Pt y@?giIeOn mRQZ (aY{j ]̎*>,Ϥbt\)md#ϒNdz|</^= EN^cS%SL`Д $CYGѥlE[7 m3 P|"֢)$ad#L*sLZpﴦmv֢xUT<>7Ɔ?mozr'{֠J c_^q)O f!?GEq{ʻ7ؚa2 GZ3xq$Xf"N KN:+QKY^=q\bl;CQE ɮR-cs'FV` PI{g-ӽa|aoX_zjGw/+2ݧ;giyZ>|pq1bec"&DE/:b/ګEbL&Gž*;a|o#*ȷ%m*ch0[-TE'$*mD_q:kTvo1x~0`>:=d)&oVɠ#ehd6(WX!3dF5OIe_ $YFFOsj!{g?lI^u{ӏzD#UMab)aqbR dGv!(Xa!b CƤQ _*[crH\J̤%Z6NIwi;'~ԑ|Փbic//~q$jf})fJl+`.yv݀@19]q8cSiǩG$ظx I5ٖ7d?>STʛIgӨf*~k_UcHƧh2ؠD^Y]R|TݜxUtsr)RP}R(vnEJJYYdba=`x@&Eh$?d,:Ț\ʈHԳ!=&OzRYƓ@Q^ѩ ZVPhC"zzAg\ HoF>&hFfY }mw큯^/?OncU/oHӖ-?bt~O/3N^GE[yN]+0vLr4r47N#4i14wew=s|9^l8g}Enuzk\ΖSĚYC3GM˨F^/hNTӍq*f_O7y^`ot5@D v:u&ŞP L Yib}m'sJ23Xa,COS!> j1* JeU,a4 6Iy5m6Dq!9HaEsF.YagZ3g=:{QSEQHp6ZUӻ.U}8 ٴ^gP|\/ì$ީ\cv0iVfkLnFq  {͋{3h|>Kuq6fьHtӵNl/zi#lp{u=?es%(3L\͹_˯;EYq`0xIDa?ݻ^w%,Ĭ"7ҜY YfSl2n17QCT`kQg1s;ه^̾FbZL>r#Z1ONf_MO)j75 rqN5ײ+ @H1HHphLyLQ}EX ;q6xu/!2eJEN{r>/ȖqW:dEgEgoC!I׃OWGmFuxSL?/ꨊ (R "Rb R86l[$Q$; whڴ-h/%h~7OrbbOu+j:_Z>+ +(TK^=dud4/,mج;Uhۭe6ؠ.=xkV;Q6@' E`$UVg$$it;umքu.5?n҇T;Q>v&]`e-׺ɇ\Gf$[G5]^GDw͸"h3h )YLc9H;lR dNaӪ2-r˾9=ڋ.jEJ'^e_}`oݥČ{f|g[pTV.^OkhmN`8~̽GU=jn7fV{vh 1h-ފbڱǥqA"MO{ N' g6O/^//]G?/8 _E,aSsYzFp wx'|훦Wzo =ZYK0X`"k"QKQE57V"FMsFV;|zI\\aa4fd/]eО\GraҗI_إXҗ&n /M\.=/}>KP#İGW]Ŏ[U2\%.\=7yT7۷Ւ7_ FG[2Sk$H3osA#SQ\AA!+7݄"XV%sOƍ|ꎒ|!xBy1c*ygUNB(kB.0Z\3ruT:a-XKԚ6.`%'?,Cg(V&e}?ӯY?Z{1{4yEM@vY啜g? u878iA\bVTbVfgC_Wa& 0T7ZT0NFfγ7|O}mKjONUYNvll,z Kd'n ;O ٺAhO7 Y ^ Z\§D$:8A}аbbWƦXA76reV 1wt_&lp$<&v^5KUya1h؋Ko4.ءǃ^ vWyiy[Ƣc\~?S6p!YY" #,5Kޓ m^dRJxjTJD!IQN)6ZcBL0&wF$JifB^rS:Qs-9xdM\: y!o3)3tn tNcʠv2E4M$0icjp7^I| Bq0f߿~} 㗿.˸+^u%%/}b9ufs ]DwfJ*2w^y-Ft TP۲MYZJew8:sT_dA(=78<ìhs%8K5waC'v:chPrЉKOTDchc迬;4|)vŬ\E-_$<=Rnn.,78X~3 j9h?ZM|漦% $֤%By'R\JM޼R֤-qKvQAZ#2а!< C{|1(Mb+2|&+Y iV;)?&~gktଲBobv\8=@F^n1W$w=<@mt xuZ88NNQQ)m=ȴVʳW&nLkKiӼs~@{> hB 4%X P$1X&A 0y7\@ ho!IWNS`am'BApfބD[.&dK}6&uIW-ҤOu;O6L7?Z;"B` .n%WK"Op ~p7,NUlhF㤖`,euz9M/rb9 siN)^sl2*a30$;lN|[>KK⒊6?lV EΑRGj#vƏ WqK 6XR)\\m+z k?v W[K !\͵:&hXW][޺J\k\}Z3ΏaNUbc!W@'9GDcra͎\ \%n%?tJ\*uWoU3\Fbt4UvqZ+gWoD8^_+>KZTMZT6_d^qQ햸7ҷS[,4IUP}x3/8Br+d0Z,, @liڬ % 1.؟j Nnkz' zVh*ɮX N:+|gIU+띢ͪrJk"tpgh2 Tq*UE^+XnQV-hD[jY2[el&,ު{sT":&:eBdc|Qb1JK,oy`̌wBZ͋/$9^"cF1Ӡ 0rƋ@  3/&eTxAv0`ޠ^| Dc"CZE2ƒ - )g$\̀0O-0o`;-0u I3jX QrR@yUNu(cI} y,(xXi(ʍ[ *tk+Y {S4c@(qH(ʨqƂ  :Č  0|@HgW:`RR"QQ@IDRyIz(U `3/djD&@H/H DTP()4Jb $Z0aƠ ~l9u@a iACxQ/{nG1 .Ybr(N2FebvN!'8?n9@*t ]:%ZLU:Xcc5FM0Gg):` Bk=)x@iA2QY`$ (gLtnX2Ѹyd dKineq *p4of,XNe~T}2y";*bA@f#HIac0t:[?ڼ13 c`A̾ \{#@ztɍN Kx1* 7Db&: .e\ !JE@ @ڙR{!UB`dUq2Ӊ蒤XnBIQY 4o1QGÙ+̀5, yQ6Hꭷ2^p`FAc-Jw3 abq؀$>/ &JPaR)'UKV^ac n:8dᙟvm|9WWfU=?Id0ue0M%L{&cvaR`am˿4m{Bk1>z[Es 0F;yp31F0x1ao(Gxu  0;]9Jޅj>= ILl#Grg.#*P P`@H $]@VHz>Hh@v oܪWɰXG>G lE]("S; ',$rmJDDy[0% !*-FJ xT%#.R1C Wa,cle?{Lڰ b"U Ҩ,e}e ZFz@.sJ\d= I3d* Tceu~5([R։`7TR %q$F!6R FxX@aܨ,%FZ&@-1E zȕ^B)Q2`G&Q`uJHCIH`~5"i07lQ*=jCUP\j]5 lcY 4YO$ H?@ElI5&Ϋ\YGjwW$"!8m5c #zWݬ̯| #|7[A< `OuGo9_MV]:+wly|{R]ٴq$d=\}cR@` Gs?Zs@S@}E (mpLS@5TS@5TS@5TS@5TS@5TS@5TS@5TS@5TS@5TS@5TS@5TS@5TS@5TS@5TS@5TS@5TS@5TS@5TS@5TS@5TS@5TS@5TS@5ԗq0XG,$Kh@V^ +hJQ H)))))))))))))))))))))))U׹GL hG sy4J Xkg) %{[c.`)ͣIW\hZ a%qW,].V׃PqY~>?ٳvD8ݟLiv r^_UiB+/t@۫|p?ع`7edS`bsԇ*x//<9N;~<.^;t9jMΑۓ'n?V?ܣ(^oQW&[2l/tip8$nzCA{]F yG z.6 !wl&S{Ej?@_ϺhG~2Q~#}2SZ冉\Ĕ椔q7u۱{v7)Z=b1r|*E g+]Y#x79^~ v_Z?˳}z뗃rO?Nyw\r3{Lm_,Oͮǘ`v6zn$V?o^|pTڌ~58YO $!RA]N?O}savڭ-:z=Z \ESnSO"P{ ^oQ1)SGcC4(v8@uELGvQ+LTrQ5熗Q|! .N0[)8ФOw#d&b2IՖ;aυQO%%_NU!Tzَ&u=Ɖ=bVٞvoptoBڢ{s=H~ Yq 837iF |ZԽSá$mo%7͡{#I;,5Fufi3כW$.E t~)Qx~m)Hd6_,Oh&H|խ_'ֳs[^O^;;_kY^^k] :9KH<|7]IBHgU *N+}r}{cNwrr͵.KƷ2\*[5LNjfq}Za0ǛQ%'.c]S9 nɜ9#seU2S̩t Ng&۽,l^w)"[w$]TZuuxӥ1*_}Jp0N{{&\f6L멠I)CGT;_RDf0EFsԋ>ju* :9UUфWCt&L+yv7rui@wxErruu[}?#wX|]*TW?һ?/AiW*WvutZOKNG|8i ՚N8WrqJW|>Ke GrXgw;pWG<5'+x=u)c@.Z[cY$JN$ FP%q${ ~T͗3k[TĄ"fY9޲%ZW -lY}I?mhe]f?D_Õ^贴CK;|1kBe~i~wCP:n޵5Yd$4p6N*UIv+k(|I ) o@̌xNv A]KĠZj^D (% (x섌i8Kf'Y !_tYbCԀd .SMe%H(o&!2~PI P(JhZ(%rs3.&qH88T迨J =o{ލW@xps~du3z7`??Eyۀh^#Z*(|CwJG6/WS֋f?$ BEpIj#53- w=~O>O ѰeVfc@/_AK6lxaI@u{.YI9L3]2P6 ,o_Z fb `؍~ K adR uA/1 9Ŀ}r각8ɥɥa[TH#YzYR|URKv/yz_^2иxWow}ýu&:zFyNR|1J/=t3[fPAwnF/U56k._]kzרaocmmӫQ=xy3@}9m)$1sGXņ_ ;0KS<įvB/ N_$̰ Wz3T̒/5lhgMS3#Xq/CJ4M;5aF Mӗ SzS aⴟ5@V%)m8Ѡѩ 3eMʗ°5|3VKB/K9kghTܢ۸&Lk}be82|J*xeaѰ{}Yʗav|12@/Y77YsCə9}!+"9W+4KBgؽêwNvVHOw3l8$C<8Ka|JeۂXSpɥB/QkNa,y|1qȡKOLgc!. YB~d\U"lٓ2pCO-&b&%8LX1&)[<yf!dx2,TV /Sٓ% h,!싓|,!PxB/ftiv.^,K7uҴūC/^jR:#&;AFiS.8HQeZ~ _6 /k6IZAB5s__igŚJ*b@NVȉ$EJX-&bi3JJ"uaEH̳ 1ʴSFQٗC~`c.)qvl8]?CvW*/;\V7ncj#$' ,$L3dӂf&K E(QEi+ۜ="}tf5QKВx^e"[~n\~0CW1w210CbA4D3X U̟F!/kq_(JT9bW+Z%PX\6x.KaR1K7%> €T,/j!3^ȊîZp_ư%VA6$.csk\Xbf]A~ƛjI \%,Zb%CÆR6Vy >&VB(L5IV Y!48l+55ʐjuv1?0dXKrhsÆY8L5KmOۿqJb`*eH'i3"I!S9S)/ R{ۿ6KRQڨxTy5.]fꍤS)Lxٝ^;S*M_-T{hQyd/J qQt'΂_< *6o7Ì6>FgkfwWcAfj_ߌ..|?j֞/Tz]4^6~;V$pH(RH/fWL.xdBx&')Oס2ȗފ:beNf5eOZ PH&P\ya-)foZggUwZmj6Q H0}=0 '7m>)a#k g@=:'_ ΋9h6k?M.f'dEӰD/!+=zefiĐv*qp, CE0Z1Fy8A gbT̀xlSŰ2ln%NPeǒrJc ,271옫8ö㤍BaL 9ECh8а'dBڸ/pCRyS^24C=z9$Z>nh &'s\y0Ht,WbтJ<<))%8OS(犊$i"05a>zq g2a-+b60l,zL6𚉕8Y8ye:n%d,oq+Ɇ.4PGa}IV|Qa?[c,}VtfŌ@l.9!ܙ=)Ba7PJ@d]XY-" ʱcAOdXq@z3j^M#\k& Vtuw_*f.ToݨXgyQ1z^ܳ64H+d=P-U7YKpZe4;ʲD}zw&+Jsqыi!(7.V tJUo&APQq߫qui[HT QۗWَ/Hi! E]…1dwTUի_؉j՘ܕԺBB} Z  ѮQD>ye '{nUKe)IN y}W~^(8Ao˕dda)ʛ;:и+ss_QTV Շ?hKpl|W bt&t 標N}::ҿ]v\I*?V%?Y֨}_zTwN>gkd~Y.UW*h(8DFcq,z*%|7>IW?o<|dV2ӵ\vt-N^>^̋(K3,сNݐɏQ/ >$m ig%Ia1rq/G'UF}w$˗dS{Y|gf>?+R'+ u(R/|1ǟ-eB2 O6d$gƇ ͠ [L5|xmk- ԟf$4_.oϋKr[˴=Du7E\Tt^y4+Jg6Y Y=3?m lDeaHph䡳c˖s/,[ ވ8<\Ϸx&]y\t= HG(Ɖ7J%]¦72*^1]VRT  1l o>Z<;h cȍ'H}Һw")yݩ 'N&wp8j| CC[ov lݹwH?ج]k ~4י<.㊦n5D9.f בO=%0&SwB< N݊fKf$O,-h,XpҮ#uTy9G t:OcP]D2ηcb3kԷpnHډ8+QPLzI`9&Hϱt":0DcDS'9wP$N#җ.wytb)Gn&d .eߜz2"\&mⶻ5k6POЩM'xsGiִ9( &/J]M(^w `J}Hă"̠UUv87z> *RsnYj+HDiM*/,{}. a6rokayJ/$:]GfLf o-] 1yOݔDj(cm}IDQoNX7 CQCh*a Azu?NqJdRB'Q{:PQBeJ% 0nd**Qځ)Ĺ?# A5#JEO= PߜӇ9A:0жDdZ|_nd))S MeM?IQmh4@#)_-jgg(F]}eYgw2〠 %([NOaAX/- ^zme y#%<Qz+>v@LCo?.&[ jZ[[4cҍd]n/&uj]D(#NܬS ɘAٕcub<)YKia Nt@1^5 sB4mn);u5p*0wmIjt"cq,xUgpLDWqHA%/ց$a=<$ӯFYr'QWi[z 5SՕ+:[gs'PM|LnN#=]:~&8&-wS4M:q@zɻƠֽh1~e2_5Zwd$<᪈>Nt)zcr0D(t>ձX¢@ JGۅSVcK48NE8cŞnjǯi(u{#5"Ҕ",E^:d6-^`*\y`2N"DGy/8],4,Mp>)KI*#ʀEA%E"&C\&F/}kՄbu9mg)`}o BF"壔b%$Bq^$޵6c~yA蝧`!r؎RKŹжdS&)ْ!IR &9~ s!W Yt`$a,#40j,Z.r.1 _9f90wZK Y0T̽:&Vbh'Y\Ce7 hX*#.E&-q]~ߙ'[Ͽr-X:4%rO[ [azo)Xɳ h4<@$4F{Rj=@/zPW8$UV(h+2f,4؇̅1/nl!Fy/Gn.O4JHrl,OȥIFuAA?;΁V=vZi8AUmZK'S18Uzb" RMX@*OagXM[xt!y\o g%1u!ZG)-*f]9ͺ#^\$oAJ.6'GJQfZv9d9,84RwM!INaGiI !E64}/O=D)ʕPĜiJ2$G)( )zM. ՟3u] wAeqELwH`̡DE.Xs ˌ(cc!-i x 4F%0Jwt=|E).L+dnr%SPL(keBޚw^W ]W<5}Hcy"kI .S]y@kQ^qVSѺʺ8Zj{׻J[O}Nw;c]W&<O ý \m,[8f 4iҡV4Qs0[9-*L@°y'(I  i:o097 YDIc.yRΧNH6:]:MYZh]T«|8 L\^戓9r&r2Սh1D/7dv'ҺnkRKcT3NzT#wv ԗn"fut 4]'3p<lmSvNOcl*ߝ-Ln'ClU8IȃZcOxy[id+z(Vow`T?%_}cCQ2J/tjqL^3`ud̕?`D+ͧTCpb``\;@]L?qL#į18\© 1T 5]ͥe6ˇ|8rDnK2'yy6;3jIͨs_U^[&"cj`Tuacrם\mR<y;yF22is⛆\C`$ᤶ v!ASܢ@cZ=ìmke[Ny}Q3fhX1mGL)ܢ9"9 d@Ȼ Tpȳ֖'hbL'+J0enzsQ੄Sճf{ w&qai3@BE#3N?qNMr!w.ra nK*پ}G%mJsԳ}{A+YRWj1+m|ƚ̙0wԄ (RR7_cn.|^ u~;q򑼇+d`NSW,U99%{{#$qNWkz\WftׯEM|^߫/Z ~7k_YԊLhP7?oؘVUf6׆?7e UX͏rAInϛ"?@(/GX^C-[M{-Qlz#N "j'->8Ӱ>8qitqR~\cm *o k^pr$CB"g{Hg ؜3L#aty^S.U~{֭0jH`l֙E;a8,R{>+}q)ƃWZQ춊,8۪b_ ͔2.Dc 0᭰LW@6\] }|RŠ_FF+,"2{2E ~c>;`.׹aJR* ؋W0DqiY ~눲ŕȔ(D*(!sUW ^1,=\,hw4\ 6=,mU^,kUK،VbwyXeJ3$y)Ql ufSb`!+!pPݾ#D@pSÆ!]6ylkPgtyy.oI8+\2QЌ ߿ޕ !hxRXf&ď'KE xHpJn]`ZPx|zkylͥũ:#iCPhע {nn?F}yJQ8x-oH[!l=6gEJW1*I!:_?T#kDFT@N+]۟v6:۪;JM2_Z\HP%8\Qxg >0L!RɪyY<[1io"EpVwk>{o.T&$B>D1sݤ\ 2~xC+6Uj,ߕ&P &Egg%lKΥחISV?_Y?s݆X+%b` ^&v^3 RjREa%, ZxFmVz<MpaMxJي#WoʽG*U]H`{UyJ)V9=BAv`s퇕!U.6;t54(0JQ5Ǩ Y*r>?nc$SkkAi ASϠNR 2 Tc)Rx'zY22! p,l\hbC(9/wrjV]_:~|/32B+Ej5PT+ E%D4U740L)p.vnd7g 0tr8jBBG$p@f\Ce 1F*v)Nыr;i^x}5ܪܺRVm%rG7%RH VƊ *|7Gm)/pM贐qSLI(ґKesWN0`_';+:KsD ic@@KiLd3qhck`QJ"oHHzd'MZ} ZsNQeJ"rtQPف'"B Vx.LO$u/(J~ow@cƮb]4=\W\-˭f?xWP0:4g6s"k:WR_8tzq8|ӱ TR!y|{-Y'T& Hx+}d8P2I}hQؚZYk4wEaHǷY9C\a A4}.,[O^K K6Iph=AC1XvKE+6 :b,z\G}RNu/:]jNg|g˶(w9;wUE/UJBu ?@cv3V\۬k3/O Fz0VYD$ y:ơWH0 C'n}ʼ`4|ߔ*wѲUi7waA 4>}VE+XZcM쵞;t//8Zm6ۀr>KgmNÃaSk?k,=36:9Jt)N,Џ$*r"^Vm9@c|(ˇE'0Ay߁cx-Q2 #+/Dpʷnig4!gl]ΨhmcND`n֑ 5;I@]U2C}Q $^%] <+W:ϭj)S߫%JbT.0/4y1y q nOŅ:S$Um~Vsڣ>aPl\OMX#xYS9PCԣ cNzCkDN +|zFF(ճJ0UyIWb%b8`jەZ(j44; C9Û=+%p,.EzwJxnʒv@FMX϶4TȂCq)(4u#2 >;acӤ2#B{bqHW,ڧ2L2*@淓T fa@긾3ǐ1HBݔ [47ZUuciVz-{ARW}.@ۭKl>봸j<3;)ߤƠ3P8avT>NVxUO޻C҂e-[3>+cС~ꌹ=xg\gQ f5^KY86J`߰g3Sn3ڀR2JS&ď'E h͑ M7>7[aԼ7f`7k+[iӏgՊn]^="Jӑ/!*7$,C3=#g+X#|mĖZ6E >iYdhqv]y;5,NɼYӹtXJҌ3. /0d*lk^pr$.E!fm6N43y'a|\Cs&+I,4(hE{]MBb'mYc]`q M{OPWKQ$EReE&I3Dw֫q..Pļ#v`L2Q[OP[oa{?\6okFpF( Hn6[`<>NJVW!zq$M}Vx; u'uO ’y4be3t;0TzQw?xi.h sxO!: z)zS; nt7))0.a;lh׳a퍆lI 6zzlϯrBUj(p>ߝMuB C`HT!'3yDabvԶv&g*5 YbZ=i#p."qz3F #B sP':mbgv0H%W(>_" MFOʦd^>Zb<>|,DBtUg 9n15{0ޛb(iI u̍{7,| %az)>g.ka{b.DKXE@M铗`hoBBDk~8^ߜ~Jc*6C$;>*"I ^Vfw4vcuj'd+hzۘ@ rRԨr"w&/3s\:(T(zD$Z7ڨ 9.YU#輚=5[i+1VFo|]jZj|?ykڛ:~5l -fpqm-)<"Ь2k&4-0ط]f$v{,i Zr-997c#_HW#M7Fjx={]2~5KƛLE9S+h^ (k fX#Rđxz=gXvgXoeXR{n'gѽI {1:ߛ~ds-0ޏפ,KO}q֋ǷyʔȻTZ*]ui{XQXc8/_޽;ɘ/Z[7Ҵ>+3ݫEDqge% (JFaǀ $% ED.W]Сkh-lLS]vdeXRbb1qu-lNcc-AMp$Q- ux(֦֗<:C0ѬVWE2$NEΩEMKɎKB .?$/\3Y VQw3 ̸/Hp;[bV.^lBILhTuWh.^,E։2z -.Ypm3EI ܟvɊq%6mR|xq"ޞ鑒qb:ީwz$S[E]\jIKaԆlO3(iiGW(nm}wG\̈́8#SⶑԎ h&XI* xV\Vɒwi{؎~'ejה)_IvM y=i~8%{vMs(Gmâ**/:ݦe[_\p Nw༩MŠe _h[L˩qVd 9e5sM wsQF3LqeIO7吆SFJ xTp5^]dyd'DtE{ʳIJLK ]fd-+簾93QbY"Fݙ=4'IJ{ugt쮇a3/Sƿ~Tjb 1,Ѵd4 o{Ѭ7G췞l<|r>'ț!'s~q|C)O3~,U8??>-Ŝ"1iп> O#(tOHI15=F4#SmߎQpKs>ӊܒvPʱT ݇ -30L>O;?(MRII-t~s2wޟoI/E{?=acL?a\SiifmT!pŌW$sJv`ǻQX@y6ՏWL7!r2 OяN1ۣGw+F~=??/#Z?2LD !J2Y\ymI@d2Q`= m<^*^f))I i4fauXT,Ĉ R8r f*' UeP wiUBM}À?^s0ʘa]ɨt!g6rK{"ˤ4ק޶DfOWB`T9!:zFE=rIq%j49?}:ϧgKI ?";%hyۏJ?&}e` R<^!mẀ(Ea m)ϊֆN3yAtW x"ƃ^j=8H IVAQcμi8sӟMsQ9U>F$ ui%#CLTLyib67=3qqt7 `=`:3z_\?}9 /1o ?zQX`QT4MMieT?O0>ali0f`zBIPeRz%~d2(쇟|XޮmI]*m#%\l;=C[qVzg`Ǿ5I/Lҍbo;y[?coПzxp!Zl40ؾ$)uH¦=27p4`ߜZ~頺 J2UFs(t5_ m\ #*Z# ,[OK+g{;\ ࿿@$XXt 5 ȫ=IvYRXw$) R2A Cq2Jp䵦VDCh RZ !$sl2T55yϓ,|v4h<>[E%(^!AN8yN킴& q^y. : ޽_qs9O*PnWij~S_SU.veХ;vK[7e]yFvo⊩ԹMO9.pj)Nrh\"4,S3pg.tz˦85{-%(5Eю7Vҍtj0uXFƘaY? 11xu r;$6L)A[`_(KMM6lzp3wZ;&##E0.DmQ|YV%:g&&PIhnȐgVPFa6~*`&1m XŃ3!6*c4p4%삚!( S=$0$#9ua+g9JP2M}PioRFmv *MZwj.@I4DyeCZsdhIj =ȟPU9-e&zKۖ\qs3stZ.O!,q60اYXIqRʟeR Yjs/ɱƁ84{u)=1!,{TbJ=;1 o"\xrojh$8־:_t( &!CpK=&q'1%HZ,#"U*Z|v80^ WBj6ԣx\JDG)c PcS m[EfTc];9E?-a>UމjT9a axaˮnw$Kd^JUnċޭeo߿ORܛs~^f|>>Dzq/\eVӔYy2+O뙕I !XعTS- qm$NHS#%Q<qL}RiU.O3;8t3a g~~\2W6lmuyȌZVi8M~QWdB (ijBqQBg4 9Ơ"R\+2';7kl! :3hؚbZ{=3P.jS!1@72xw48r" AJ1^e՝d8S#Ea `|LJjî*p !zp)x0*k0Ra:1il.ı1nti>SV 4kS:analF6Mz%aU^b. j$hA!T9)-u(6X:؜?|S>X Cq&CJ(tv1:qF^iY|pNSi*>8s VQ1x^aUdXOm߱<Mc^`\@Yy.U|N&۷Ww ]@ԛԕKC2iS(+\û*t l*gښ+ձF6xcȔa+PZV vٻ&qW2d*/Cj#<>RR[yЖB )CeUxxaJ[+y}2&e3čuH{.l5Ւ4_&PJ+)%݈u2' ) vh SK@)( R0Ͱ&7hp4V L Q*C3"3$ 4sl%G %@Ϝ$*.u]R*3kL4VD1S.p"1w|`I9'iJ*1s]iocNЊY*{q7KԢD!JSLsĉH%n(ؠLfRۉ!9}9v2>]xq-MUP3*;Sz4Z`MFha0;#q@rn򻄝O?Q:wJ@7q-ۦ@Fo~ }Вҝ<76~- !2 %52rbSV_nGm8_ϟNOuyu>*FpP z2oqۉF< ,~>.çjd8zJ ʿ7Vp = ꓂Q+ÿ_^x?kBWw:YpIcm(}kp̿?ڦ@6vC|S꧇@pW4nd?wzxo cj^f$œԌNәY$~[ #Pv^~*L9}/*,f?9 îP ;VsJU:J};r1SuQ\6nS,sn+ܷ~sX|A)Jr!S(+rkh%oHGl0xʏrۙja擟8lmț$pOSL6Z%4Β$)X$Rhj0MHyD2S`KӘq"Ku(88)@= U0] EJo~堆RkS(5q K(8:AǍ]Xnp0z1brTŠav; ĉNZOl``^,#v7Q1  C{5\3I-f\gjHrهy0([ċM:a MysIeYt̄ dp|BuZR)z1XǣLj<]XFl5.E7Ԧ'*JЄXgí3f'ZdȔ0 ld촩]kRE7JT&!>$|oє Ws[/9&vpkZ5eV 2lvF447 a-}Kv<¤ҟx-[C7kvQ6ʒ&bAV0Hd 8r|RHVL%TVE!yJ[ 35+(˘Ġ^" Sa ȇs p8O/f{I˲o V 79UaΥ6bI*Q@ƇԚbk2X u(BU#@ xHhCu Ƒoc&q.;ה (csN_p6|*FP[ 9og)'_9}X|kߧ=s~77<]bHLCع yu6VEYXN_A( goza+ r>(1/(?@GBܰ,Ѫĭ蘇/qS5ɻ(cdSMq'ףѺou4p1ƚ/QMaUu;ܮ, 9֩4n`.&M|uO)0X6O@rn9u{kKsNNZi} wZDb)QӱN0 ikNl10I*G.Ժ֢]ܙ6-+,ZXZs[] Q яHh JbTF:RBȺI5EetpחQIlr8Zр`o ;IRaae?ѻ'`OmNijB~q$N|ذrP/acvra:VR Iˢ›PMoth:_毁9 dGJv YE9[}m/&pnV0;jO*̯.rѐaA"5KW`Y7N=/M"Wg^5'3fyVE0xn_bppk{xs/MV œ/s2 ɦ{1W'u:R;$ԅp@c,poqh^M'0:r®Hƙ96`d/?y>yg1nnXh~NkZaʚSQh}I-E"kaY^*!pճVJ{:b6=U_I(M8=XI(kIlR56cDSm=uT㗑0cs 5%HJ=qɦ."I@G7LK. D]dJ\ h&l.XOlph*Olb-m+[, D. `xXc 8Eko!JxW췞uO>-\iApTQ4E. d GL3Q9VNYm[;nI*h1s4oSGH+m^ 6[Mӹ+['T ێmETCl]ތi4:+0-DPccuNX@^!2c^98>N?L0j@6|b@R)z1XH1"OVˈk<^>&(T CĄnl]>yMWt2.&r8m%r2jRpo*:V{ZmN4fA LA~a FtSG`ьuD\[ Rr)^;O6Ѩl]>}|Q}TQ1M9tT"W}~ê1lL?lm{3֘OEo5:!XPw#sK^$ಬ"O *=FRkv)I*dRx2>15yMt@qՔ([{}].qse*OB{&Z</x3ld$p "e4_#l}zSbd[ %.M[El5M=-,Kو(=o?M Jk> _'ֈO1b^mnLCTI*'L> ~00kdW\!72/vJd3{g =P'ˮsCʥo3϶biftie)X~}ud(ԃMOO^A)oيz*pSj oFĄN&?՟pubCbJ&O~~jzaBj$Ffu[Q}rU]Yo9+^vؔy#Hz{4Ѝv7awadɒ*`*,eu%`XReV2AFOЛg,+Κ"nZÞ+q;72#OSoqzcign)Z~F>5y! h}zc9]|,͚1bYW!#Vt8 SJ1"' WC鈮X9nhk(нOTS?ߞ/N} Iju1m;x097xTzUp<랗Vr\x/d։Ԡ‰K!hn=J\Qer:}URrͧwl-Z^ɥ5cmzq!gf6l99 ˬ0BɇdMc0xm:dh99u+3sHpCuLȺ4%() IB.X(g /<|\K־C!}4ne??p-6ܠTyLI v;%N+鱔8uePC#rԋ.eP7Vٟ:tȪPkJ.O(B);y段rz zYB҆K9a%$xLg#,ܥ )Ϡ~UKAV `_W!ƍ.s,7dVRpJ"{ņR,PSOV]M /\G]bH}&~:9S1nEiih|n)ȶMAy?ӷ9)o/ZFkfm$7Ӊbcg8]@NV=oz"U=ARN]Z#J=B6!l)pOcdTci: S1]*#fY<{&ݛM3mԦyc& ?=ZwTv8c0Uh۾,k}5zql 08,rl0-fx㾦iw !x{qܞq}Av DVb/0|eJmD5c<#6aR;)ugFJ5c-)u4vن]NNY3#>AeFhM&ȋF`Z #H;&;Sf{S5c<چ* .w _}BDh [:R,'p mJ{H)z]9!I 6f:m"6vf"::\HfGVO srd!]:,fɮمDMPBsJhBLF+q 槊EY8'|-/)OgC%|U6-!^sgnC$ zsh$P`~EkX=v_uQs/&ymM3^6YS? \SQ! ;V 3c'$N9D1eM؋7MM.&iHP@BQP)lS%# M]("S EQ.diX/I"ΊߦHe}(z@ z QئjBh|.XTh] 330RBc3fٺF+0 Qxe%fݏfȖ,~qUImDj>8PD9ӗG9gv BQ:']%j-n~2X=6 k lY7Rpe4u~!R,)syj>w'ҺP .TMe~'L[b3¶}P9x6IȒN7 < , èi@07@34&_t]8l6Xq`Ӹ;й8MHAI=2N9om}hL ݥ}~B|{94Ŝ,޲Uh哤dO֚clD$潒 BYVci;PLjWEڑ^P>}Acx'3:Jؾ,ɚ9oR7,X'sv(eS"dEl)$_ieb!i T-csʅm*% VT:) D_XG%b (f*#9mQ;T[F +e:{M['Ԛ5E kBN "z$%b@+yH FaLXƠVNƌ 'Y%(lJoI.BUxgͤk #MH 岼36 y]˃ 1'*ƭ,J"Sd|c()<屛v5[dҁ-J dMzg”sP,;k@t2L}ZWF7\;YvVT9+=ӘZ-D|t3d^?rSoB1:KƉɸo=˖$kEۚBLBh;|39$&8ᄶ |NpgQB '֋>t^C`|<3? N8dWI:4'MpP/VIdsŦU/Vᖝ՚ݙ g,a^{?$77 Mzepo0:RrCx@?y;JҎؑ%x۬[&/}If!xN\a_,w^wL?[۹>fQvtZ!S㳠]6>YڐmHgbk02b"a+h _@?D )X ʵn >h5Ac{I{WױJ03uS "><0AШ՛ؖKkՒY>IHgaGHL3@EG (*^n[]6dջjt kT|1fK)U 8G>qFdm|m}ӯڬnl 3e и=jbw΢ԮcP0*JW *R.5ɂqYgSޙ Թ fqjbe2A( ~(nLXP6evQrIsӳиb*kFroW4raxYl˫1_3bf|zw^sQ)GD^$K8M\AR!Q-Y%Kž'y3j(b(tX9y3pn}-]Ԩsy:kuv5ɹ-;#[!X m.ַ:EBQ9S:s\8r C/Q̊ pCG̬WGcí*/6rh,?X ME8hkH<Qyd}E[2x#-$ՓhSֿ..؎P+ =+pr ДRUYYݣ*b/ y 8I` d|sؑZQ[2 aN]LlڎS)!{ ~H-)pp:f2XEhE*GMĄrQ|n痓l?zz @2|&\]1`)=Xc'6;gjcF-\:M\MZvϰ d8f+HFK=Œ ֆk%q$I QKU'KEkg#r-Ǿ4*Z}N%ݼes0h\_TdZ^픎)k ]T(QA$^R0 *! Y/KFH") 'T({=)ZoŖ[&sx=@ zZ RgZ"{cmu]2{Y~4W# KuǏW݁;ޱfyur!˯.\0c\SkD٥F`TyMLJƭB n Xqg'4߽浼[өWK݅>E󿯎.. ~Bwsg~C?T32}W\A8vVwu:7j>ݿz$|Ԃ~{_JRε=8\[Cˡ;UBmf}/dsu֟t Z!/.GD'j*X˷ΠBt 􆏽RMEz# Z%u`󯷺ex(/a t-Sdv.XvdIiK) ҏAþ4({޽Aׇdܘced-pT ny.@4 )- /9~p!/dKTmL~8 )<\=ℰam*/&*COP1V!uzq` uqsY^m݄^2.nyY]4~nѳu,5n*I+HXl]S9#1YF]w{jgeR2VRc3Xr]y셦F]^|F;gUz [x\rŭUŊ\% a];()J[µCjHI E6a 3\J-`(1{)혽))B{.os FD>9H_ڍэ} o^`;oyqm>PrYY t)YJn3}ўlMKo D10)8qɩG:i?Rm]->F|j_bꜥ nKC]r+-j)*6l Ahٛʟ? |Sa ym3)ў(k⊭1M@P!/&B/FmYjdeq+O`p<`06& n=$/*Qe cTy9KIooX^/gR}Z^_]JJVWPa>v.V9+#0sYя}}==rU?ڱĉX햐:F{{S\il5={mCxDݾo9f*nr ;Uڋg>W^<xkwEY75jgpSwAw@c}dC3~EqZ^uQUum4;lP`d]#Y%EJzU6uE)d ɱ5>Xr컳-S5`jgMq +dڋP2.gK箓X1ْŎ%Dp7 (ʗ]etnh?e?C8| `k Djyx>YzY}ڶ$ƚ3_|<GS<_tϤCj&5"]lt&Xy D4n]_*7}l'PEM=;=1OWC}(sэAln׉([øǭMqx&O @Tt#Y1ԻE_|%6ם>^p%^o1#, c6ГG߇O#u>.msI'vi!.q_\rU6{[I+iPu);P|!SVz&$="g[KL}kQ|:{?X;Bχbs_Ţpʼnª໏%ӝ9jCJ^R[C ]'囬 *E)+JU ,B}aw!RXn :؍n -f$㍜=vCg꘱Unj:fl1[uv:;em}Ӵޕp0[}h)V+8"aXzբ%q#l1eQV7Nz2Lw_Bs?u@nWѶm TR[8K tgNNf$Clcw:ʨ3uʅn)sΊlޘX MK~0 ijV:` LY%1|tsluA"gdzwqn{GDEWaШL&&#UZT=vgAZUA/B! DyNDI8JvPBVCRAÌ<'<{`]ƘZ5B$ƒMTtF]$[}2|MPYׇiiK5&ϾbUwi<0cSUމGxJrwBj3yș7Ֆ`Y.cF Dža)A(5RA_Yqօm_l|Kt/=5!;O?Rjy8#;&c0Z ۍ) M9:5&RmH&QI*5sT${h)X]) x8"ɖNM9"k_dC4Hvz8jسNb]eܭH?dqWuDl*zHPfXA;zoTvi {u5Yv*a5J>6bQs4q;dJ;ʔFcQs41cShnFsw3r,##̳+1 fY0ĉc]d=;`FѲeN;d+t{t=LvʴM1ӎ4vuTRef*)A%Z:WH)=(KC.i"#P)-6Oٍ'x%{>{LE071lt}Uu h> M$zkBgZ㺑8H2og1̦'&eӲd$h>R/)hwEE"8'*U I9i(ybRc,q`8D ;RY$՞CFI-6kN7lb) <K+G=R )p5Jh:3 Н<20sz8 < ~i6at'WfrM(?oH|5Q<8\tL.Z'冄Be)L5ұuHQ<RB=QZ($$~#6(+NȬ2.ƻy+ߌe2; &^Kc4b XSN>|_2oDuqSTMQ)s l^*C6ے >]ۍoR4ۻo0MIif eH9:&IقiЙ(^-[f3oQjZZ `4d bf2[Aˉ;[NG^x9eC4*ҹق[+>R!X-,Q`{L}@Ir)7i' ˵G.BTs#cjT4l0p4d-$MKuQWo%k0M֬fYwt B\YM1>c:FyHa_!GNJ-!Xj}P5TDu5PЦ! yL/X$A'{`.-fi!7K YZm:bx_T}k%7TPrSAm*(tpvZݰPa}Y0[mb Д0 l霓T\rg=ϤqISWbiR ׎>^7EjĢ˟[:jF1yɡC'$Ē5QY.1C&,U=|m>ZǒY-žV#45ܨ>*jial!(n6^ϜJ]l 2@PxdK=ևo,$!b-3!LS.֭#okyUƘs3-;46-R-ҳavҕ: խ QQKCFz@Dy?J`dS|\C`{vM^ F1aU8U xwiݐ \8W:f_l qI?˂b-P㼘-JE[64+m5"N:6JH[yP^ wh ngj0HS9#,>ffnOJ)%Ozڃ^fp$nr)H{C0!Io̸gÂ]VL1N}|դwIcECҥ=yFd7#TJ 'My,">30,zh~<ͅ["l$B^?V6\m84z?¨YAbZhp5s38â̡Z~j15wx$(D+g~~Q5[\iG-".9(jv:S4NJ-܊DG% +H !R  Z?ll=]zWGoZ=!5&rvΎT-زOSw~ᛷ)n_8T5Eve{'(5:U/{'ڴh˱q"O~r97!Amf&_385/ŦIl6f|64qC[^آǰޗjA-sV@*uo>^ o!iܬf*ƛUىIT$kaM2 oToǬUw$$- \/6] O||= SQ~ijC.x 7ŸK[o $b~fO?_M/_>xD18xB&/ i}jm\lgg E'{]!u&/hj#uvc}py[`_~Z>hQeg~G)c ݬ6Gs";.mrIPI c=#F2ZCS򹅞rY Z|R5%9SY@cmHuqkM4lWfh7+DY!mV> .Čd#ZN?Ml#:aģ[v6ی"_K>.p1Vd}+nއk6TCحxlvbuG*z[)󤨶HiDQ(C4cLkQWHQoE qIp@Z07톹yW ^;#%Nّh8?np^V4 ?6pעg. VzMңӱE,ʂ6'4OWvOMopDݏrlV'L1^сfnEv0 ׄ,IE< Vhw8 hgáV sbV]4B0yYX<{h)@zHLYZ 5k@"U^7>ͲNc>zZ 7JBo0*Xy& Y>ڭH2sSN.TH_>+TFl)w(Ζ&p ȥ1NuV# 9SUkH骸( bk4CRݕ$=Ϧ_M+PI;Z55!ǐAb -x׳:(4XYn4y!&<+D)UƲ.3jyÐΘzc A@I3\ dݒX$FixDqz/ޏ^^"o$rvsfF@r?M}l`•'y%hC)YūE/bl@*4s&q]:x5ˀњo7K $,A< CJЉazc^0öغo%0`8QA KCG+3vIJ2;mvȒGvTC,,XI6alTj+qf)TM{bxp9Sx2˵1q12Ԭ6s0Ǖ -a#87bn4\&Ze]dc[ŸT @e Yq?hp*o FL0|gc^,7g@{t#=ðv}p!8_pnLf7-^ Y>Y4C{cvϏk3FA'jbSM-\2`$C 궨`(ba:(Q ÞX*AU<׋_6PAĆ8{?mM4׸Q}6O1X,ڳ^i4\= fhZ:fE(ȱ`ZEETT8pܸexDd-Dir0)櫲gAg֐{8*y$ע'zMGrzx) ʼn< =8֑|Sbcw! ra8]IA_H$/E0ػLM`4n[YL߯zq[jj˖-xDu)TY0yF8S蛔gF YH$Et>H/1H])(6P""sVhuУTCcٰQk~~7ёcwWR όn+=?R|;~dy3YF6e/>kwV2^s0W7p>+g<7p-.XMIN0nuWy'g)EÛ!j9$t֩_mۘTmQ*|R/T yzF}Jwŧ,ޏ-طayW߾ZHVO,[PkߓiXE'~vra(7F9k@t%fO*YGa;ܚep@%2]m}+9PU/)Z(YPCö=4HIt< 5A?S݊@1٘7Uk#m*acثX';(UJv<ե4ؒF5aC:Nhu *&0;4큏 .S^?Tiؘm#S]6y#Lp ڮLN`_06Ivƽٸ0F>d\&oǸĉ@6eV1¸l0H+=t9Z=!Пҡevmp}e|)y&F> ꁅe71j XӀ5=f)gx@N܊ݼF#}-<`d&obo!DZ-w faddb:VN2XqMRDz-)VNi7!hһA)3C̢ś0,fm0-Țik;efFŎ.$x!};L;M4VԽN: ,SHC"IdMDC RHw_JT#0GR2'j3)PڽrjGjD}@+#Ei0ςdvmXhDڑ:T-Jvq shX|E"-E&jׄo kvFpHN89l$j4"J7a0ŭH.QJPJMa4(l|t/JRr+bpmɼtr]dL7s:]`18y`K^jM+\ #,֒wf-Q @j\F "8TA!@˜' *(G[V),}! )yNA,!VFyO`˝3,)bSor HAle鱰.Qyqɘ*r,!1"16ݜؗ5Ĥh+$(v=JK5ՄoA-3iZzkZNnO_O@\ԎΠ"DAv59vP;Ɏ-mkixǐCЧɂTZjG`8IE*aҨґD0.LS;4bV5`P䣉=}Ƣ<^ü}A^p}jI7.Ooӓ\{ _vd\d7WW7V!F'7z*YV46vCcYjTͨ2%ֻT"kLaCӯY3kϰ(c^̖![{BW)jnFTGgTQ\M>rj}+jף,ULKcکh;ބZԥN΋+LaAfrKB)Jx0c^|.<كř_./W֡gG١ք8f0{ws>6=T/l}ϟ@:wQd_ݾ NJGGGcX70G5K;JxKiKDUHpIo"H&ҁKɢ0%u"\>帮1+LzE\fw?_+yC.YE'seٟS >ͺ ]wL*3i\Q \H#}pϵRZGh/Bv8ެ0> <{hJg~ իh+MtbɮݩY˘*I`a/q!g) da3,1alyJuhV~{BQܺa6㍨2u6*jjԂJwϵģs|=Iꬭ~ir6 @ $(KfB yraD9aNn!T%yV챿Vi"ޝ=\6XA'[ MԱnڣQ^.ڻL)e;Bho&`Yp?itQ]'I'+`v ĞƦ9n} we$ ھ<8Lkվp!X v\M'X<˃Rzɒk'jq'r{"nfBٲ,YAFb.sAR ϥ*9+=0^䒾Z>Ihk)|+<ʈA R~1A9M.ty!BʼH* SHûįztޗ{Y@.rZZ28 &4H/1p%!XpGoBqz֝xK;]f\d\*.ih$21l<_'V̫@n[Eöc$y=19,;pݤBrzAw*:ϖ > !ٲp-%SLsov<]SUT;Wsx{yIc6ӈR2[O9c3|uvmsvC60 9]%/kJ+/܅<8fx92 hҴ4Yb'1rx#bT>xhTd!貅 Iȍ\8J+2Dl bGd4!Zğ/Ǣ)HBE1߼Oˆ(/:OmJRG.G??۾Kg_TN׾\X#Ko=߿̒k d1ΜMs # TQmu 6h2Qf&܍d""=YHbDzVH7/Ƅ%9>xU  h)dS0yqe5[J'NfuK" .)m >Vi.X0(QUEV9G˷f%P@mnguɬ$w"yߑ覲OٷRiC26O8{c Mg%WJY=q֍c~Ju؆~%]3䎎}Ѳ"2bW &WH Yi.g3m.y,߷E>^|}kwrG3ŪwF wĢEUfrem$SE"novz:tz=99>H?{Ƒ XlE9a'@L"$e ߷z)pT$9.XQ+, $fL|KaB qrFqUvtԿ-XvϚ?*),PޏBX2fʹ& A03>Ɋ:|L)hs/hc*9M<)&+@-Cp֘\p%d0VQ`#b#Ut64%yX NnZk柍0$1|9)ʍTaT#ƴ[ǭ@fA6pY+1ﲼ{ZٍOg[F)+lĝ4"Ԃ)gp572#ynTjSMia# Dd{bΥ66jf6hScL{J\9+Ͻ5i@ z4X%ot6RdU2I}lSn5rLpRi"W"mUtidV` E]XJ $-3CN,!S)= RPҤ0P\~(U 0/U:Ґ;r)DÊ(}A` 89|&AVy c/9 h*tIuyAv5FfғLizF&B'Pk8r<F2Mo.1ѱ\@̔l Lu"xiv83QoQoi4kV"D_F#^ |LJ?R4(u :mg(mAM+p~#TUB᎜vTwJ#wf ]"H;%cV M39 b7\>9L5K";}aD4D¹" c6'HUʟѡ|2C2c߭%q J|){wUXNtwt@nD.3iPec]x8W-(RckӍH_Uz{xߤP@ ;s)wJ 6*(D-|7ހL|Y3<[fLlIy6iخ7Ji wnɿ]5g;,{=YwQGdb 82wOEFOw{f:܎@D{~G?_4`lHH^s c}a@F6b,$DLT}8ֈ2Rq46G~21Hɪ}0Z@:omҼ`*FZȤl_[-n"_G*M4d}LMQ̏ޫu?$Bph4q-`)B))ljWiNf.PE1뀽s"XkxiɸN *T\( ujg>IWw2p!RP(2lq"I)|={@[Lj\˅"VvQN85.5`~L0{,8q"J@Neq#טC%>dBtYa5'祒ER]I=9OB'/S~=[GĥϛB+!ȧa#Mݰ ԚJ#@]sդ:88x+BDdJ6?78΂1UeeCNC֨Y=r‡cD 8yB1mlqR w$90Ӈ tpL97$tvg$}``nWC2YI*y KEhi#ʶQ?Š++824.?`‹aZ>Džs&:ns&:nεv2g ̈́1SEsr:w#Wrnw.k]_q5 ПW7ɸuK)//ʱb`ϸ/f/❽_yCo|gVE ψ%.ŮeUz-#S lr"l˜bA`I S$#rFEY{μςkH8jt= )pJ( A1^'C!)1"[A`@+ł) aNcFp*z@tMH\j!BPuTg0KzsI~?FgZguۊ~ njܢ\@F{nub:iris>`>`ݨ.xG4g 1*x@3r>7J*VKԆJysm?@q65B'R)TdTcwqb+3i⼁F/5XNhQfr @9hI:$9DV,_$b]o{v@#хg =?QX|= Z)pp;w(ĐJMݳ<4!edK\G/òG%ƫk8iU"BhN`8]\K~rFv*W"(Z|" 3hz f0eW;Ϙ`$oYlx:HZXa"{3q39I|a'#?* LckUyH-1_.(a!I@1"IGi0fZ"tqB( 1pFlbXym= \!OCf+RXtM`vRKǯߐ7-x~Z -͏;ӺGYUz?E/ &ٜoYU;C3}m;9bF#x>Np"9(a۽aP.!}9kTL",Ӊ c5DlcUMOQ-kmr &@˺3ifd6MlՖ!1} 0`ɯ_;a a{lv?4IG}Lc7ه/zn`>Sc#og{<+!ue2ebH4Q__AxF9Jr2bu d!TTZt1|&8 3Dj} pl4Ō$YZM[.7.bPpnz !Jbg⹚9r `h[١Cs~ExA_=\lJ]ĦeRftk=])nA pA:ֺA:e=ԠEH}ݢ djFrHcIt؇]uš;uk-ey 8 ۅ`x;+ڧ+O.R%z@@J%y4IRRMZ Jq]`MZ^"͓pjHƘ2AgNskel3c}{뤩BǻZbwM^ɗ3Q Ҫlą.$.1B]2]8| d=@ q㴹cWX+_ZYao܎@'+G v oNGpAkY8JAkNPh5mJ^չ2u95G7&ׄJr5sp{9GZG~MVs JI٥Du ֬b+ӕX1:jt`QY,Y. b-&X(wns*~MLRPG=xgJ R&*  x=CLt&,aA3'˸V oDE GtY 쓟O&{[nNF{נZَݗ/*pt|qYĮ&)=6`oQ?/ON߮>>VMҔ.' f;_5o1Ѿi(Jhρ*xG`cgf7_oaW|`t]7;^;[tG\_O{Oחy]ys*,K}V[嵝TjkZgU#3HヘC0tx͹Fh>ӧǗx~~p_?/:ɛp mNp뷓+8  /Q{s}}չ^@n5\vR(\wJ_˿s) 2u{LwO{S{w4?5:E۽Nwv_[,QnYj\тJI/ҞՊA/]j<,5ʀnːGs O죽NfBQ~,8jn͏e6E㝿:/% :M۱gXCy[9t*(>s0;ymW~^v@_xWNoAhǨդl(1u͂?97f ~w:dzny#k,+d9k_A)?mN}YgJgeX>82kg TO| %~% `XLu/WS).c8Z%f2{eIq+<9/vtlqSސC%61 C22^*sil7ja$a<[no5)W DtrKQ(V,fO?:M q6X=΀>6W1O,1歛$oE0=F^VyC"yCZHJ^%+Jl\cKX2K^~g5d%aO6`I5DL&,e.1DgyS+{ Xj,%s)e+F*oh$oh"!G0_gYO1(Ǐ[C*ּxB<'c)vUTaoic>lw[XUX}5 mpIJҌhh*d0!Ūbv*河@1֢ {~oUYB;HI kPh@T_ՠ5/M(VYDN$!^XcUl0êH2Rx8B\g ㌗'ǝ2f N%RIA[{Z#I>%8"k$Y#In^sCb6-NSꉲ VS$ITq$\P`}DRsQc(R)E%k,Yc-_\^cKXdJo(!  0]b$ Z İЂKFb\xX2F^Vy#w-%0[H.6[Z|+" m߶׆B>#߾˰>]H9$`0/Mjaf$cJ 5QVT&1i.kwzdxh 灱x#Դlr߅2 dmh ƇZͬ#8C=Pb~]MB+ &e_Fk~>(PEnT~Ʒλdfگ/pA|XAJ%raf^px9:1SU$0V0Azv2#az|3_~#DƏtRƏ5~OzekX??f,YD$J782aSdo*GaPX~@Q/\\h!mh|C +Whvn;N܄A\cG3}qIƤJk]nB,olVlbO hh2zۨ>F`0fCP<>>f~O*MSo S^y)tjn6.3.[)Mxkhja[ҩU~'C xP>h俏o6,>k;^QJ ؤ]B娶\ x8J7phY: imw4Fr@$IeęT8M ,H+Z:vNұWCG1݂<й㴬˟:r1]]g_hJ6OAW,ۃ+/a+A1`aaaaU wĝ(1;("eŬpa!rtM)|L,@8SĜ@? E?z|\P|<bO/H b&_݋^(w5z` Fp˰wm7faFޞ|h?F2Lg͟ x>=DJjuɛf:Ac;y;8ۙ{րv!.;5sbAy<D9#Oa-ncupRN$!ʑHt4 jdg͵o6).LY,fMQF"-+_ٱ$ su0`;OQ6hBzɵS 0'¨&YD&e(“+q=s4 8Ο]HJN~)͝mVƏv{S'778̐ 6;rZ(^)`["lCfBGiJD@hA$x@%r#i bE-HIItxE"犩nqő${,1 F'y̸m%u@eCCG:BV#{dSKL:G4kàm`oHuXoŊݝoWj(09w5 9sgͩj[4'bzKK􆩭;0,zqNU{a4\w1{DQu5cDѠi 1# 5 #fdd[/- h- VD1/)(ݥv&H(ЏR n4^07L%643L45Y HgE<1~d@:%@*80_=@X -)iЇczdJ h<"p钓$+idaaޢBJh杒qLd;% ۩XȨ>-6ӏ \3eeR`'=cuxJoJPW/#2㣚#+'-Bh2dt@&Gw@H$Wjs\,kkSܱEm}>>ش_Noz*:&Yzm4)ڝ HP.'xά@CI&tOx8:]>Lp*"WܲJmNԱIyFv_cn Ù%ݳSH Q,8j@ՠ\@2Ώf4\žrRm<y_08& CrӚmv4:X^iJ0VX he"(hw hMPbULE}n'])"#kKȀWR{x8ũb;K3TOO5-7iD `OjT aw P*eNUgkLȑ+V쉲Q{P7lh#o7&>"vad],nc-y,\0 2dNV_4% [S d]ٺ-b@BmĘڵޚ)|؞Xugdq[VA؁UV2G"$ĜڏFP4Ugu[3agfJmQfN"I1wFJu:CUr ]uLaLaopiI~R署]}K"ŵJ(u?6"uÕ F;FL^gu%$rFDrQb~Vd%qݖ2 T_gYyJHPpHB lc(knש;ʌ0Z[=+i6E͝z55ͯ.)NtҵvƄ:qhL; R29/Դ]DS"2/d^ۮO\nX%G LQPğu{@ *г:-G+m#INLy0݃NovtCȫlc)v{0}#YYI3H ۔řQq*' 1gm hn76:ErN-輅pf!g{ݬ&|[\irؚɺ! vwsFuVJ[y/KUbTrsMTh\I=![jD&k $`g ̕8Hz˼vW9#݌i)drݿLW02af9LV $RO&YwECj1W~8mu0do8cLzsEj3S \Ea/aL+ p­c17+ gF~`\*L'..gc$yx2N'?58KrODh#b `vs=D]W8 Z/XZQz GHJʸ킘p=}ל+o[!NK߀m4Dk(GR%QGv68Z:YKU8%U uӣ0~kZ3pOڴԵ$2J]&ZS3jk~/S;k݃g!M\ =M莳4Ɲ7Ww|?}AKjԕZl!JyG-J{kz#yd22IdI#-q<;m1~dV-eB`XVzK077otğ@aR6 d)T_5NG?\_?>u-+>˨z.f|Of1oP[(nƗ]P[D=R_[|(> |W拎nPCJ17ۋ*^;oyD;&6W, {u##ض4G #/+CdSB.!68K)8K^1!V Β58;D2 ѹ<-$boIqvʶvNvYCd?3+%u'=Kt" &qx 2M1KoOS;DV:(wh'A&2!3|جæɇ&R}k9Vah%}>|{0Cܪ$!}Չ˲-qLK6K.1\R56E0-SPzӲV'lyß!~#MiQp߈v%Yʵ/^KWt`D]D["[==)<=,s2SCq) D#*ő g8%S}r[,%D(|懻A= Ef^Cf[{>Z)K)O*i5AypaĖA%%rc%2X\(ؽ(ΟV u)x Α8pé扳*\PNJmTws4lhgkf$X7`%0Bv3:KcƴDD'+ [zل(ɡ [?ީb: BR=uµd{`r;:jYռh" +z8M?_u;&j=kK5'R/8e4( Z yÂץTd̚d/5%FQ.)zcU)$Ǩ\Y=uC^Zͽ sݲݺoE[|6^XI/afKKʾZcksnc2Ѣay|,瑗'ats? 8귓wٳg,O)gj*i `^ te^w#\Ζ/arIlqSqCzYZszhNrOެz}J!9A[/J__LPd@naVV|fMOV"j9?qWj)%޼2wwof5$5pl(qE\h<./aVLoHs@zY:,_Reb<4]KXDת"uaY%\yIqQv kH,EuZv uH {C%y\ЀtxKHP\]Bk:eE(ac⻁qO#",K9'ԑm@f l+Ug1p5U4gf@bCC62@%b t3ƮeG<"MNBM^4d5),Xu!)hpz*ee6ۨXk 2̬5찼`)E$6AшJIlr`>fȨZn9*Dj}5UC4#O.Qb7{ Xˎ)(0AD\Я: j{&2ZU%VޭFI_8| !+OuފK/4B(-)ұ`Oy\zKT)̔#%aZrLhFo.-J3z;`}ثJGb2ٳgلjsړ&4l Hfֱ3ǫ>*xᇥwSgB`1m+qY,)yrd\p[Orz Q]kCS#\XKQ`܉B(cXðK= [M 6B[(7SŘx/oBGj8NƟ݇MRh)R.5 C<e"Kʏ5R*%H$- p%"\pj *ӡ$Iv*68(cޗ\Z.rK@s=##Ot[PH>]FׅQL!iKK完zKQr9 XCp`ؚ949G$XB}%`H@%! rNQWɺQIM/}~\>t_613$7^͚E*cbRG䬟Ϣ#_|}^wc'WexI^xxDu9*wL " P3NGћ_\Ls~@6,fFbO;cs7~.O߹kx>NpunKVqۣ?OP ϡg]0J~炎Tk!!^hNRdLo˘*3Gc:1M"i;cMΌ"+mOdFJj,=hw`b/,c,Jc!e=2&tڃ^\;I O%Bb'w,( [p5ɢb[C:ӷ:46lBڕPiƣ=~h_6X$w=n$s~հu"HLvNk44ֲvӃ 5frH+5B)pyǂ)5Ӻǽ$;9"Xߏى'?EZ)ǗX՜Q㲙j4`,t،]s7Wr+x?XKΞk7+j ErI934H$0J\E5FHpרhwGhv`xRŖnXt%.YKXww\@X]|b}q%e4|uD/#ͷg5V7 ʾp ʾW/ghzFd!;|HpN\켌ssg5@0^^hW7bTl>9AKf|B "beFO"b|dQgU9ǓTH֞kփM#Q+ۛCX%k'~%k/r3.>ώucrCYt$Ǧ->]gl7kZ >1FYzR [":=Tj'zm[X-3{%>aO!`۳.?:Kwgi欏օ&m]PkJ#4o1?wsc{= 6Ռ3'cSN_&n59101oqmfm")p6EkXG kݠp?Qh?ΐRLyӭkQBnѣn]nF:[ i~\8 {y/?%C6͹ɺ;\96(CTÕVCM2GEnF@UL8<(X'L8<(+Z 6!;Y 5ɻcȱـ1^T.px>6+cs=KeogŮLx`z(}0X@o'^iNkCFJx{ۺ$/zU,c e_=GFSۓifB[XǘE &]s#Z lx H y,r *)(o:&DHGр>6B ki#5r" h g Hր @gXxkAź}ٗ²_ j-;EP~Yv+uPpE0Hb7<\>2!`EfJbpe2J s cAYv1('Ʀ)/Jh~A;:Ǒ:pX@SBcW2Pg"c .&ڜ x_JOkp8*6O{L'&)D9? X`YYdˀZǓ'u=7Mb č,wg%׳3FLa+k-B"Dy[m"=  >'I E@A-{Ԃ?pߟ,C멽5w6N}4@ΐ0-{.82"V)70Z"GH)sR_]'r-4TK2$L,+Fkt*"#5Ҙ듊%y06߸As,4gfR̦r,CR/IsCŮk<˘-L4`d1&($d}ģhwnJi.ڤȕMDBKmARkԍNjYF}3_W@WZcrJClPCk I*8Ռ&XNQSmNDB>2(;KCXR ,N̰IyAfc)Q0 qKqSbxB{ ׂCG&uF}PAL& AL@, x1am:-lupdJ0U4Fm jw:B-i[ 2 6 $㩹BwkmeҊ@5Ϫ:Ҍ!,XVND5`Z(#\\p.tH7'NP{M1[Ni-4}9pƋo!|~7IfG=[=|X>!67So_zwc+$[LOJc}.gty93*] &1~;$ziոof&@[?Xq3'BoΓWß'Gg;?9MmTE| !wN֟jz26f1qOi<-W"*2% aAڅDޣ?9mME'8@(n'wl;9|Qi[j^:v9mx 9Pv,?m۵ Y;O G7'o( N<%ԲWkȖk#֮;N'5aK޾0k >$\3LF#Y0s$o~_ˆF+g_8!SSccQ^5N=Ե%#p`ж%N @^Q6X =pӁC uyx7 ^bA=x=ql: TFSKXnV-JS+hf*d2͍­K:Dsq)_8t#u]è)N?;R?X6Ɔ0R.b\3Ql^qI ϓ|}WXLQMsX}X37|;8y])'Y鷯_b `>ia|tȌ f 3JƈO?o c>߃4|~:FW=ҿӏܽ&$Y>^n}eߋ˔p1Ԗ"'R"%6gR JYʬ]ci#8SFPWL0qU|~[ [WTOWߟOf`NÕ?XC&P?ez5e}e* er'6WJ9' #J>G5ϝ4ħa:Jǧ҇w7f>2!e@ݒKoe^Mg#뿲XHvYdE0vĸO`G${J#i[3ьv;]*zM^yE>0Zo/ֈV.q~ ,y_~/__WKx*±TlN@%c@UAαD 垊Vտ772U**Z jU?Պo$)%k9Os3rexw!_?U\yųwcߋur޼݊>pńB>% ȶJ9Ccܹ*6t$&knVq<ͩ=}wzћ";~b-z6y3okvʠKG?M:)1ni^)ws#5r~믧_/K5irOߊ}  _6xOރzz7^\揷g/M3D(:2A CaM %$om,zc2x7%hpf'IpxBu` 172eh"$i*zgjޖN`*һ)P]͔y7.CKN $Ϧ iקT7NP&M#eXwm޹ف%d E [q 2Kqdԟp튰;* QP8b5Pj  H&N*kM7'P?BߟK^Ӊw/V8}_P_MݤYݞFv¢͹vh&à;3H&([8&YI wtv9y[hK^T4rgi3TUٚkd~~ ׊ɲ]+4PS lXۃ*tHt^jΞkΨ`s5'8Mgi6gn)|\ ??9/C\>E@<G`n!!./1`Q~cK)\-) :͘ 'xbj T3:aE" V6RvY.z Q[eIk[Sy0瓇2~^OG'g>Cvr66ٙtТ#[s,:l=Pq/3t˃C-xteѝНa b,OR' ca1Q!p/,hX̨ЅbÓbbFe7xcvjMw_n?}cOuۃ{~njn>#w|zc=x=x.ߧk֧&P֧tC}Cg]{3F G"`/xh]m~0}yii\]?vO XݬG?3{uU8Ić= 5V&. *$[L2Z.J  ewc{xs =y˲![6K΃BziIzAouMEo?\^s;fh|;⛦xqyşOήd$ ^>0xKۋ)Yy;ipH'lm=q(KOG)5̚hDRh8h-:.t`i-:. AO,`VdxE۸W~Zl{/G9)n-ξ\v}z9v&^ ¡e`wi-r:wo|Qf{0GL+,ony~Na3=ƴgk>75.zu׃h3yqI̸%#oy^'D8?7h({p5Up _ϰ?~@ _F\7zWrhk|}?e-OYOYxS; ⽧śkoyS8!k+8oUJ1hITKQd[l,R輪)TL4wM|w(72ay|1Q!@D^{hyt?c,~sQgikJ LJ3>[58)q+tH,"eGͣ2O:qY3 //kfiL˚Ò2#$6ckxYZHq~"È[xAyAs<ScuJGj 3PP}<^m̓sMB*jQՠ.A7ӊc^++bqFN2`Ʀ8nܢTABPNĪXY#̶lZm5)^}oEz_U1Xq|s~t_mWq_]E7}O_^wQ/9o֐5ࢲʶX`Zri RU-*Kh)ziNLL[qTt9F6y \Y;7֬%iپhAf+@2 Ǫ7>AA[RÁ5ϝ'r(8&`M"r7t6+ح{a2HGH@Kxn7媒]4y3RήOu^eXB Y}=8BJIJoD(})`rUlo.d`@͜Ʋ$lapPqdc/@]QwPo5ed4'9btWf<ʍ!'JDjcn5e:2q%dfEZl P2EX70 P2Y걘3 V<8DYϛ&-4Jgd@MRZHZP8J-0c:!c}/s'lke*U#SXZͰu@6$R9+œGt5Q&H$r3pPdmb2h,f^gj( JlW'W4TJ2̀k hx>X{2TcI19`N5GBj:Ag tAz;uM;ʇ ^``RaLfPC/p׭,C:yݻjX| Vd  8`vV AMrOQbR$892jܠLTճ T@8g/.gJ4JƬkZfZM^#8,4P6[{ћ jVJ p1;; ZYL A()C#F A V$LtaXh^Ƚ/# I/)nMG͠ ۀ8JG U'2YЩ ֏U^׀\żҶBMfF(NA6ڽ}WiƽyQ'sUъO|Of_oFQ"Gށ]Ŭ:"sp]/ 7#0Bdm A` D"=YM 9HP LJuX:t*RIvVh# ERə%{q~/ jV䪔T-v>\@˘K2|ȖU߯|-  lVDOzz{zӄqy" (;%D!0a! :* 0faYV` Ć"Vໃ{##b>w?B,xDUj)cNq`0"XB e&Y'?#]Nl$x <3kS R\(6d ^^b ֿJ]&CTc/SoJ0]ȝw<,m 'VT =vҋo>&/unr4\S0x 56`DĆ`1=)Pnlaa[%a(Ijfa05l431ytU7Щ!ab8KX);9CrζcrI 5>"6d=9 ځ0pq):UQ 0=#r]3̃6Vp`|<@++hROL8/:r WU҂;1 b& ;ȂH8QіRW=)@ׁZ%-/m 4l:8jMJ T J-Z*]*#&`[{&^bz@( TD@]HbX竍AԥDiB@j¾w`Na;U5OT px? +`!AàuZp ^=Е}!=&`S0%h,˻,ٶXe%a<Ve/,42.g7սK!^0 A--4dlu[-r[~ m[ h-fg!J)2ٵE0d8ɭR:Xua-X$ , (΢4)b}2iTt84C_X U(7T)0ȜM>(W9cak,_Ttg[~QWwVYuzpyu$ҳ2_ j(KyքB `st y_j-+_Ԟw)VYo}UYƜ.]i #{ֻ 3Ԟw)@%m͝+aGF,&n_oD;dr:3CZ9yFBJ~~|BL dD#7sv 'UKd3y{#7*_pw~jz'셂N$h DBg7.^D~0yϗ׃Gr27>c?z׽'{x!u?qؙ~ķo n}zv'#o;KOxy_nf`dzp}Ԑɛy`=z6B7ߟ_ F|νi lOQwU`>3p #pQM%_"36ӿ(߾㞟tAoM.R\v aʿX%C{;%um&&90gd-vyAae?df?9v=v.)g֊ޛӅ 9iҗ@u6PA,| $ 6d T+BGCStiklDIۼ Pp 5U!t25v!Ctc ',@ ZOI4j-ue)M\8ȭ IV̩Hʗ)FND&T&@051kB)kׄ ٺ5m%@WmV!TfX6}UF6slB`֤|& [96mxF 6st✢3 @L*b хv_{谭|_u@۔3wsw˚B1>y4zUc52W/K'ŹjOfi #`9Jh&lA^ٳ>jq|4 "<)oO &/+kk6lX}?&fo p6p{%ʑh8'-<~LlRh[1{j] *%H 1NkƟB)"Dž{G?Xc&^D1&M\)^{s8^Q ԫA5ێA im֟$ȑje^r82 Js9<ȫM@Cuw5JIfHvO=r07fOt{:销EuS}ݨ_x%P LW3` /J?I? [$O*^XBihc SX>u%$n]k)= Xw UW\%,%fmN,t9P.J6x*e&G<}n3 "*ᓇ۩RyDboNanAYl| #J,$t[Za5ʃ\AjUz78L49ȊyDtb`8a4 BpI^~N7;2kc="xMo`lz<٬׳a~͹r w'(R p^odR;!Ȼa~KGѲ #<  $mLjr|X3bsXoYFP=%ɥ']ۗXA9ߡFBr=]Ɯ]цeH(| PklI?p^Z}YUZ]%7.6H8ZgEo>E_䞐,fKbXozӨ7 cu*Ip,qɥS@rA}UKuJX=f,B&L_d.e.2y 01- n/(67g0_Nnwws^1riQ쩳1H A(N"FLab ,(q_0`L|y Jo΂OGs&u|/bnO>M!QMV4 `.O?z*dQ/GMt*P13כƣag{~r9+8o~y,EDVRBǞ Y˾u6A\ 9d1wS+"[j4 z'#Y2b&c Eyh< A ~:`KGb+ Hajl3/ced֪KɤF#h-<WY?A` CB~UƠEd&66b'~?(; |&cw4gip:.EPf4 N#Ȕwf&AuRt7n+Sz:v2\sF)}tgMB/.$CCܪ0n6^,)Zׂ|, nd;N6Y*1M_s+x?G,i<NpHC|P5W`-30ԥCҌVvNZBx?ϓ d6;W 4;0 g;w3Yg&>++iٻ0?e0:wf\-7=A+XL9Ѡ 37L1b4?lx!|?,3%jFpN|(_*@lh8ӧtLcRF%`!9vY ܓ}1:*{x>!nBͺf^9 ͈`lN4La*߲< ^04ˤ2x)(Ӫy1Z}>HM I*.~,-5"x#]ڜ+!0u`Fl cB,s!b=(QTWWx\bfc݄IlE5aElp76[CZ)F׫ͱL%A͆~l)0Vv|쑅թbbGzJAdaJ xdyfS=Aڔ %27 ev$D|ύʸufz0~KgD *H:TM(:~0`gn/͚~mi ,x_@ǏUJ-]S]4% S+dK Q,l\p'Z&G<z&;.OW.GI B֜&t:no->d4vpU92)J'·)y&)V4K PcF䏅)0}}{ O pX|6TQax1;]jJHK&QZMSto>"A'/5dG Ƹ8_`jW'F)B _Sj1NV}YBf]%r(V,8%;s~W;@hNKӄ>yu^"\#e6`Wq6]B,ˀJ߾|.WWĘ&ΟFOړ`1!DR=dMOJàb '۳KXdK ZGutS=ZERռJa,Gfȉ=>$B8UҶ#s'`6Owx6åhj@+ĭ'SCֈ0ܶjRA70̃.fV%1W>2^udըea<]Dnrz,ǽ+Mg!/Aye1"X()u=h80S6ʹy!a:rtT*y`X83;tX#7&bkt~4NR'\5峺4MൺAǻs庭2㈪9n~=yKHu!)_ 16޲G Z!8_Ց-FJСwY(nOSsJͥt}|lmNU\9lfk`UK>PgXJI)Gl$CtRTK~1cp zB9y_&NR(% eAe{Lj2 FCs;TO VZxn~!:SF|*ؽ߇I\FMn  ԃkˀLeV= qg v?(g:qy165bh48gR;.6~a8]IXGI״#J2UdӐU'Ɯ#t浔D 7_s"-G-֒԰ck癀ua*urxwKD/!)wJeJoameM Ġ$qw`WPBc9օpk#p1C `'aրyfo5TazƆck/rr-0X[WmU#6sA1R]\dgww5hЈJ߅Sd$T3 T FX\ J I)'M=m T3E& Mh8Z\2!jg3RY0H;`(Gvwrq"FdX;'.<%~ɤ~U!F<{53c'POي&×K)X5iLWh,{y fPw1K 2ǷE#QX0`kn[cbEΚ wizy:I`Js#W4;%!/3s˄aXpg/A\;;RXs"TcYvJ٠SѮIPOElW8d#K!? &qٗ(m ټ$[D/ۭ [џ|6|;韖0c03YOio`# R J6nIu > [.C.YK*<_ڞuؓA4䛜$H?6[gŷ t &O1CyW-#m O IG:,wD@4#}T?=0j`b$#'$6GKS8]7# ?D=#ӳ?6dz5;m=AnjLX5fv`IeB9+ʖ))?º(3Ƴ6g[y1 S}Қ7yIڀ8~A擥 Z59/N@qR`{b9Q63 nKZm'է [V-g 3Y(gm^jD"Er݂jz Y%)NIRIxҜZP8ZJQ .[6>6:0"=˩PQ Qze:p?0σyAbbH M3J)϶|O5)EDg˚N/rkT'"N\8\%u6 o_IHZsC7;5/CdI}Å!8_?n<rS"yyw"6~62=DI7 W*<(_W hHATi9R)2EˍcXsOs摟GeYwPꑡLA4uٱ3+8|\j*6dz8*uδ*&bQ{Hܻdc%0>w{43y3ng SV4uKD v+9/K=?%rKqfi$K`_9O\8cT%i!NEMr{=|Xj77aGs<:m`%8e%0N5%]Sh&=xaFefH%oJ3Y|agLD!(^-f;Ts&xEC^]rW|˶rkr\2 `[7䜥]k4bQA_N D [?1W?/y qIkKn\Hokr բi,88=?{Oqbhp箉=dB"ϻ+yܫ%MPBs ɨ5Ca4 ̏mtBHp yvnJRA84Geg|/@.D&ƕ-r5CZr` }(_S\(.nXVa[΂(ُ%r:h ٓ3ByO+#Hi5;r?Њ]ڮvYnu÷?'&'7vz"s*fGbq#dOiG:`A (>T88B#;vN5{JI&Ӄ2(Jn w\qҏy{轭<  dL12ÄQJ 0 !9U _y^dMMn:S 2BI`%Bil" 2"RNk gHdZ;^ze7 ł,ADzga?x,+4%JPרX5+7gcc"1 v*hRRAqF[)Eraţm|/ԟU?.IK"e1YqpXAuNtgl+-Խwiu^'^&auJnv(iH1?nkh`~ Zw<ػmdWT~92U~Hn2Kr\Ȓ$Ofv+,Q@RTƖ@__pitLӳ j)P 7Rڔj[o+|_gdBcf 6p5p3/07fR(8p齤[ܧܧNVt2Ͻt?fKwocNh觨"g>N{fHYAf[CCwJ =v1CrRuj,e]..>‹ET M|\l=t]gTC";) _V kEɖLeh"k 4bq~( {bPšv>oh<*!{;~W]xڅwͮ dO!UB~3bjALuaV 9ZxgIkDYd rs^2W lgW~~7"]vLGQ.` ;i\p7烿V[>oW%};X5zt-<nCo~J=GR"nљ<3&##rҿjЙzyO~ٮBG`5.lKU LXm7;^Դ2bNL͂Y.3Ȫc`2}x/r~Yo6x!e^|;>|xpМ-?=920)щcȄ@II d<#1qY_pU8_=U>~ Q5&*P!@L~qy8`t*GaMbnd bnݫa1n4ap1?ˊE,&3.-WpŒt=s vB,)5 ǂ5UBh>Dɑ˂}qԈo2$iN3D ARh&0 Sf]oJC2n%f;>Ct ( T"EΠ9,pOq>J>2*qj:wF] ]މ>}qHV3X&JdM睨l`ybobTr{όLGgֆhznF߮B/'y:#gg޸C)iW#I [6,hLǩ`IJ~ٓ$O{ҳ%ٷv^rQ6! y"%S7[ ?|Ԅqv])s}Ψyj}J,B6:>刚D(n)@҄q),8j.E֙KB{}1̭_paon 滫(!"7)MաzZ67^e3y#u1¤̛b{so5>}tf[fϏI` &M"D"@Qȼ…H\C +P9*&ӯE0%{pZÔSK$o>Xg$VM|jsr, y\,P5 ;~Y_^t5KOƍ_>3ꍨ ;xޡLt  H2CM {*^Aj- j}[Qk) y?(`tH BU7JpU87dQwG@"qIQ$ӇZk&${W&v4je%EgUĊ7/v:̾_"}Bi H$Ĥ(BaɴI,,;!Yvu ACBA2c,I-M(<29$ !Ų oW[p*iF͜ ۓ*Zۯx4L:l_}<_׏$AxHfN#Kx))"n(^Rz=q [N\^@ 9ĪPUxK Z`BZ봺kK* i)׀JlZ[YC5uZ vZhkqZ~HfV`Dh8*e4'j (A@tQN2D q@N{}$ G@MX YMXND 4ӦxV4VTbv:ᵧorZ})jkfE\ ' c)f~eڭuEghaGH n؁f' :l>zk$OFF$ch}EilDnY+J2Sʓ#3ݫng]Y]\ZŒ6C)PK \(&2$af eni2q~)pSE. @"YGq4JGc ?I۳#G<̿5lK11`"&ra rA~g5o9q DhQۈ6;J!-sXˇ6 *m(sY nr+4 O R 8[wMV{e? XBZ8z?=M;vpŌ丰-xVkot))aξ$*nIʷ\v娛`zt]Y^3_#gl0 Gݨl9u5s* Q7ZeB9ҡll9ik nك  GRY&T O@j̹gAřf@rkS ܒ݀Z̄jVGǑE1)_i),R8cdn &12;j`3-QҀدZ :7g_RG-՚/ZEQ"[N`QTG*1$SntTr3f Jb(f-ibQ ʜ1eyΕVZߩd _~YO7~ ?j!Mpgz>.Zǧ~.;T0*?|]_=5;[cMW?%&kΐ_3A6K 18vy澅Iwo񙑚Ϳ_>3Y2ص762O/!:G(g=?sɥ}$R0Nay۔C#%KDQ S#Ts5OB Z[V\N霠$B0pguXr9b~~dg2}"l8[5˳jz1eE.}aJN?G| ^Yj~~刯Ys9g@M($?~K'nPҰ<'n!o(7DˆG;HC_w;)ugndq^7v/Iskh {j[77덋fnp96ͽ_L_y飝l6߻bΊ)[k /I4HZ(]Tg낯d:$H9h.AA&hkS,uMn%GuIb Չ[Dò HU2Cz؛e#,f Jx%p [|X-PS } %,'vuTFdbQ[IN*j85'$'Qk9=78^D@}p|i4H0D6,D `AD3 P!5aqyz,06YxW ?cx /;1A ݋BhHLYj̈́R5'T, "%jCC*)maj;5/0IP2 ڷ%DʆΨ׻f.w1㗂vM҂8@Uɤq% 1 BFR9%H!#yUXV]=kuB)%^!JDKɗ з7#?ۊo˛xsцu }j|V kQ%K'+H7 `x7enmgC( (QA4ekmK*3ew8|kK"8~8g-ѱvţQ*N^îx2_H҈-1bBCSZsG:i }qiBZ#229ԵCJG:(hH0- UP4+d Oqf ׎WLgj{ȑ_1將dXi'w8-vs_;c;L߯(9lIqZjI,* )XI_|Q5lR(mJVj<Cv`ITofaԞs"6AV uMͽhIH+{$&ONQ6wM8tyF !+Ũ[R9#֌@؅ VщQގEF)’Cw!%뷲$YͪT3-`䍑]zO6֢+!rb pI+WC@4^0$u8!b(1\ZCρ`aS 0Ruǫ$oW-9qlN4H9Zv\+tw?SGS{՟v`p`;-W4 pF{3zHFo'[ l}>qJX3ē&*__Q"ߒ a3^Fulh#|.M$ٔV&?)D͊KvUgdŒRt`AՔh*ʧ!0:mE1 %! 5.E89A *BQf-p:#*$Sk#bmY"α"Y6cbbP9HHs'E1E^5(u3&f@^irc {;%}8 >{C8miN=łJ00O\f|B2(N,krX#x0I!w.:Рߑ(2Up{=wQ6"ھcA,np*Q^W6Q˙Da%EFjo+uuXASخj9fb|(z;猱T7~x56oQ'3C9HH*{f6C@O 5:5Gj\ʆ@ٸh[sx/HF`H樮N2ӊ<6smL^yCZo^eS]gĦZl|jlSϘh}ڦA6~65|v;I F=X?1D#wՒQg{N&9 &ݼ?u_p֢d{P&ފņ*vnV ŅFq6_<tM'u,8",^qyӫxXĸpSUW+qv4C*zhӋƪ}Ҋ}yvQ^<%1S @z'Ћ' ZUЏK/x>ؔ l;S,6њ[;?=;0,dW5_}޲͛O5kn3|W%~48a;|`~cv"}WzL ,\Hy~Eǖ- Nhvo;[ɚ1>.Xvm7rg ճa<mx /=hP Vuv.A{Ue2t߉}s&niۖگ=XhAY^'O ?\~ڵp_E]]Is|e "s\={qɏo޿<;gv2忇7AbCy|>4һfҦZ>zhtn[<ӽmYچYy-v~Ae΂{0as<&-yc2O L8?9ȳ.ACpʏ| &,z8 c;`zJ3pڎc`. \K~~/-V/uz|J)G;zU*.!G:$翛(/;HV]}ju.6?!5b?pS~_@kPvS[-=xuK:=k[g&B&<0np 9y X[}!GӾ!Q!{u 9y Z[;[| 9zt.ǐNijU;;p6q؃ŻSHF?y Sݨ%;rdcw 0Zy~䝢ݐtWEe#HƜl:D@)k1&=d٬<"j0R1W?z0-%/zGb4VVJ CvT]5<FRJ 3!{lh'F0>Y2]˶-ΠYgSm.$i+o/>/v6CF.N%d b-rĜe U !FvqAb ԃ&0n" P!dW<kAtV^&@h=Ux&0!>[drVH>nl2&%QM Y>2&WKY@dlFD?v)_넡 YS: بd%{OPSJ gu+i䥠#͖SgfՆ#:[@: k4k,o/׿61^zKMSBQXd]ۇ}(yD] ޽?]jE.ːOnu~!̱k.G[my;/mи%?S[I ,t4csx{pz3-NEcOviFag:>閦^[(bk}SM-s%S~8Ͽ~?"7 N[1[`kҁg-$ 8{ⳀE4]^?jcC|*۠vt4K؀[2C&NF|&2./hCf=O5|d-@9j'?}|Xs7Zb3Dtp7 @4ĭvdCR<|w vQw唿*o6 qDQچ9A!7-r2okSİc}lB/\f X][s+,!52lTm*NjR*\-9-K4=I D+Jܶ ˔f[(Fp9P̪FlЇ$-[WݥOd]{zWa8CX_s8*0WqEFws.&8X!kt1z$4 3U8h< cTIijŃFOQ\ }2B 5^8.bcE \y ؝r>+[1ƳѨM59Hy7V5Q 95bsu՟*1{,m\!ç*ȝl%$gQym'mdd /m(dl&p'/fpþj99gq5?RHPi%>) r #[a%?h ޒ:qP˱[Ɵ!ha'I1Q wg8яp: f| ']`# {,4ހCH_EGʁ|L_nBjz-b5HF-6MsŤvSz,+"}QSAtF:|uvԬf'̆yePi&ī::} [n?me#gQ*e-~5ۜ%싍vC؜Z 4QtSqxPnN'C:i{RLʑ}$i;dz:~^NfW!fKή W|\x$3?OWqFz>4 ߰{tעQ=sHajEŽRk8/8/$M:q‘>xG JAW2)hͱd: efùw=erw/,@hܿVg|#eP,?]'We"Љ޷Ս^z}y}_·~Ye^x[ \|]c- O5cZ^/]TM@]T H37'jKk4D$g.a$N b^dL=Oi !+0zi*mPoD~y ߖ-F$LX&Rc$ WK,tf&1[!d@@\*"E B9Exv5cL (t>y?- -\u`jFV5zVt`$lj QIAp\VIws\Fk 9;,{0[1eh>+Ӭ#6D%c"&d g5Q7.X^:f!D0V$S&'朁!2^{"8# !͕،Zm/?wpOIRD۞YiXU#dChAvzQ.E!~1u;جqa_|I>OONl:+h.4M}x|}O$["߼z!;IϬTzAn:Wޒ cw繘9Ev}Lp \\(]*}?2¸&K3q3֕LIG:KQfΔ]3-Or땛/HW/|uQ| H4$N#)ɐqo]c ,rE‰-vNo^eRKwW/Loɮ&O_O=aFoW]|4)L'v犑]}XiiMCeR)di djGdJU`#_0F| Xt4KB,uBԎ ڒ1}g>GGfT; 4 <Icڃ%U) yhl1;t ql4}e.M7hM6TO$Vn TB];ɢV"_dۻJs# S5.d;]JBR(i ÙR I:-߽nKvR ذtWEI;`W&C[ǬLiy([4_L ^LO X) !uژ%359DS6qO# 6WCy1PR.v3V zD'gA|aJ ,&8 [8{& 4+$xu\Q/dH:lf]LR z}hH\'i Bp]JFYGqUDGW&:k5iQi]ޕ ㇰ.QA/bJd+Zc椖3*Ã`x=q]ޅ!dzGZln}&%UL tqo"`LU+LHW޺m"d nI&PZ5B{lK#UY:MQXCO Lqzyz.3ŐHpBv/[̛XPCFNQIKN'S$ @6KȜL8!4Qsxd!"(.F&c!E#fڍs:H@&4٘AdD P }Vf#~(V#tT%t>Cs7GC[sɰؙhhM.5~gYs(55փt-q>s>$`M}*^D B (!s5)?̒bIkH RH@a4$ h( F\d.%AaYOIEۀ1/CCdgnJkFIb䡑mcwMv9dJI4n=fz(?Q=+gEzf1~^0LVrw=[kиH.Ҽ0Tױo1&Yd68\vN@|eUmNPKnͳ{f!ŷ ^B~~j.HזbJOņLg<`{- X:#Q>e ;ٿ8=01qX #Źy|ۥF/p,x,5E˄ݝ/z_a8+8l F؄t}zT,vzzABM*qF(=VO=hai}LHD;o0pGv >+no6هT5_U$E,uW4FyE%;!"}̯5!)5d < \Z]̫BXbV8T]^k4#w =rҨ-.}gHeaQEqy)Ia΃"JT& (7mIQpkFX+Zp"AR省7FZ1sk9MkL >X,H׏W-O"76znT^MU%on[] ̦jomÂ7WCCe4%=NF&;\Se(*=vJMRޓLͦG.zՃJƶ'5h-+1g)G5ӠpW)!׸뮼Cנ9Ziи y.G%|gwA.QEfR'^8gVB,U]_ќ[_o4WvQEnE]tcAiAE"$.l e**`mUpNbO&ݯe5oZ׷[UmG:M1'ItIjG#Yi;dz:U<ڻ$ ˀYѐMÇ3b y'U؇q*^WE <47 b*d֦PcDXҞA+ɢ&<̍^$3bCc05l#;YXCQUmdd;9Q 'DGǥH" ir 9rn}DJ~mJwmqJz2#EI{p`y3`%A.=3=q=Q$%ɯ>օ,V^&_N/r7:>LccF@YPǸ~n1#I>#e~W|eFt>R$m5y:sV!smIS ;x%$=8yAI'O mߣ9 Z髫'Jd{IE+*}-GJTw+.m/|M I^'h;T{oÜډm5"ESi [m-EU!7Fp0͉ Z[eYo ­]r=/;0E֊#SûjUw]M92Br#w9d\j1B`)`FgUA5`* Rvv)h,>&^+R\A.yY*oe֔KႱdY̚$oC2vūU4axJ]w`ҦQgG$bV(Z+wp%aj; lٛ'߷gF$qlA&cihĪНdI#]7CdΡݪUF,0J]n$_#p](L$d_JbQrb?D%L9Z2Kޜkm|[Il֙2j_9s]m^]84NOCA&ihdI:(E%"|t`#,㩋FoMp!/ ګH&9\f4[h4WbB 9bGS>zU*[Ɓ7Y^;[0-5|m_1'Q\m`)v8Om<:WQK9k#f̧.(YY -QͨZoH|&Ά0Zے@T3U*\ C8Jb$F|GV qc`eh2ʡ@Iq`2I"'*9"4/ sІ!?&{ Br^D#~DpypLQ́0PRk0b E 6am lǥPvsb*&ch_(K &FD[UlyT QeI,'& } ' PY׈\@>H,"钴Vd"x$6Mm" ,dGX#ٹNYlb@A*CI:MflbEVCȺ9?=mٮX z54Lፂ(x}ɿ6;}3󡍂ٿ@;tmx["!~uCtZ$-(+l{iC,8f= 4+Mtpx,=Գ XiPQMPd}iy@ԣ%@ԤH7\WFm=z,&;eFS ۾g"‚znM?*V=ѥ[M@#˺wLfʰc,mҒDMv1FPOApzhqn X=cL[ipG L7:]^\*ؿ9'OTN$9WWv42ʩW5 9hWpRm@f AYKֲ">@X84"$,J`HcyPs#Il r/E+ah%W-8QΘN$F=V#z.`'e[39Α|F~Ǵ>L:)(iBk׵eΪ 7́׍nc{n7c LOzr(SZȝϳ&dSt098b`M:qKT]>*hzbI2x -lf6' ? ՚ZƀVxP'z b-X[  ԭbd`묛0tZ+8C +f79Cb$E*)Rd|ȚKg Ŵ`HRW]t.$pP OlT9I0xS. eOx6`Н+)3lN?*ɾVdPvđ+9bV69C;ÄĒwJ7a4O TV+6ฒsRN\ KVERȸҽ`Lpk*ճf<~ޑCEc~)X^ <4Р;b=9$>)C!dG8vq=4lb4jƻGC0\jY)SLlq7-LJ)I8SV8"8<6ZF ۞gZ~B4fxZHtVgu=w;f$%jm֝ݚw:6GS-cxgcOߑ f{¶dP;#ƙ@/Ms0fM=98iE5CFk1rD9f[KQ3  e{ͼQt^3 df>(+qVtP{e+oy7{g@U'ZyV 4fvaJ)9 ,y኎emK$S 8BU v.=^{5Aa;hfIlE p VIYln #Wq/|5 ZoHềï{b(I͇~P7|90Fj Qmm==᫁QnYovп4^!QGvGrV./oW~q/`ή9HNoOS{s_rg H1Syo..K7d]7g_=۳Oy rɣ}ъP7GO0?K\,&Z(Jjə 䝸ɟh6ea¬|#x o@~#"< T>|ӄ~$V_9 <ȝzƐ@vZıƔ>]׭wݖoBڍ [볛&f T.7߭3,N{QWUvH("J#ɖNI&<}?q߿$Ӄ ٯWW5hՆXIf[rd<*guJ߱.=\f%Zbzۊ!ɱWs ^ѦP[UwoơhhBPo$?[Y }d ֝7/y7@DAރhjcB`h޹\ޛ%DIIQa$@`֐EDo QR1Z ;B.:ћD9ZSkxmP <[ Ϡ_|؛_/]/UkV2vXAlg_Nn>?g+[(ڇ*.HYJ֋';V:X,!&XZ< "R M6-FD (ʼn|ssڃ2A$rļ.: 1"xBphBD&(bbuXi~rHv|\'VVEQCʂգPgjF*+%?vs 1BR,)qw5I}Wyt{ FV䜃*Qѣ)嬝 zdF:dvJx&\n%89+ou}}!DX 6͠JS$6=&!0A#Q)v 6Y=4AQhq~jBuEqIThѣ(``}X 1Ve$H &DjyY}oJ`}(2Ƭ؝BI؋dYpxPoؖŠXi^/WvZoS2Own}ŚxV.X?]7/jpw(.ם>Y)EW6ٮp!ZgD.6ӎH'j4FUeܿ6R0D ":{fAs$>-ll#][UMMA? ~ ?gOfeքת s'_@58 mƄNʨEZz ȓtY#P GqBChj >uѼަ8EӄIrdw!Z@j6DN>u/P7gO nb ~(Luçw-Iֽ3@I&^B$\\5U%rtb.TNݪQ7/dGi&&fLWu ^ص,ʓa;6 Z/כ~5vV<X,ߝ]6任unu百=_qyg}]WQ9˝?|ߥq aqSʙOoxVd!׷x\fe'T테8aÚ %Z=d(H`V%H\V͟@VJ/.Lu=~ٽNO,Oiy|Ϯg7Y3ͽ]?ВK:z?&_}Խy<7,F4i xN;a?\` yYq#du.QUT`L3Qu UOP:(Y_C&Qt]b{׷*^êpkPWssFl\hK*NG$V ,jՂN9TJu0ſ9cX/s"Pb( W".5ISQlذys3+0G>/YtUL[]r 5[v#S|ݘNx3E9l+(S?] ~킗B$Ķ)r_ 7kcwM2Z≇-%qp-Vxm7[lu> Op<=8ORUh:[CX!2&\< G#>:gŒ'[o:u95$mՠuV#*:!Fю6mJi}OѬu6W8ykOrުuև?s'}v;anL8p][XfJ-4=Z-8 T..eqGYs+[do[I6p;Yv<:aR;$[/&̇? sab18|qDtmn#e s]aAX6SGǓ,no:ۧ+TfƷ*߈y.QiE^8G6+_^\$9)me}I:ytIE۲/frk_T-c4GZl_+_N{KB['{,oNQe(VQ`GF)(;=Pxvr2׍"sSbUFA&5KG7 Ē|Faahs0ViprpǍ"bč!u17 B[*$ URqԢCH=mzthI6 G}l rזV[R)T@gMpΗ N:t t&4ij{{~l}V=GlpT]bـ9k2Ny^YwzA9ku7j뢪SpNQݛ)|}` /謚J@/eSПUҾdW@+Mjz˝=;UMhk"tdF]sUgWĉs2$pUy_{O M3 9(@M%]֞u)9l^&Tyc[5&V9}rj/ X9&|X_x"YC g'X+T?tXC?>4fS6ff - {|o@B[ϑFk46ω!)@7W=t}Es,1(٠aJK9\H Ɣ )F<+Y5 %+Uxedy cY4 :OV /~sz[.{s{Y{O<~uUweRWF?UX~=*XZ}&8` 7`s8~FY/ (*RT `_VseMtn_fZh`zG_*L5aX_X!#WH۾)L:\h/vAsBZ5? l0o` TI)CtIBsը@i=ȍ^λe쬏%,+˹Ӟ2ÚBjgAYZI0ƅo,@A|4@kFRQk"NiHLV*n^xPVJ‹\ngg!?`=Jjg {cB9ਥ 0@1~جyhp9P3p  ͖ FS,Wab]OaJO\׽h6O)$P.g/\-ĴsŹo?wQ~4#P "ŌɵVP"R܂:pdں Ґ~փY)1Dʔ nPL8AșX[,ވ xL@ VHJtجcl:,tB|l98j<k#x x18"AR]aZh%K1 \25B0T{4-XFnQl= SkI$b  ڡLJ0b!d ¥) g;[-ʬ@$S!604F1#F qm02330x> cAؼD`Sd92Pp|,)4`>PP ҚӇ[ ~.!KƆAj(IqMキ*S5;68;Fڱ:dF9F`ˮi0ȃI0Hz-m,:V=fzsր9_8Yʶ.vk(v_O`g~vm~f& tz &.Na>%zϽX&-NᳯIzmFN{BqAhsaePX/?LtR2 ;"PDGʙ8Kc2;) @:B|J BbHA{'#ae$t}. pUT^:;w`#JFϖR\{ g;r>o/AķF|(;- !v#S3t2O/{42K%f~\bQj'zr˃A"+x1Żsi03X4C5:сPaI(GIW&hZnWq_PVYeokgF&A:@K9 KJ%kz^/׽@ƤPF\]dđh.fp0wf$=+~t>J!ؼ^1M>r$ չm)L&am~IGx8Es 3@|g}- H|&{L ,k9D -]mKq]dZ)~xւ6m>}96-2b[e((ET rycg|n~dZBVƠ]R/DJ;6ד\nk "9R_ݪsϫrG5 2;,{BJBb*WOT;_LkZ._W]**rOB"R8ǁj,IFZ "RVPDcL}!?2֎TݯADywr++ 6.s-Rz* q8v,8s`C#r}lH 5 Jo3T] q"b@L#4 И9X&AGȊ0!:H^?u1c@ci{he q f?dGY6.?2ux>tRSto*1h[Pu̳zu~,qP0zjf.[:?&t>D?ӿx~~;?U7>-P8IB"ZE0{(/cjcݎ(Z,߸v[ J2EQ=Gڍ!IݪbPFtv;f@[nvCB"!SL )bE]g+Vak QJn}_]]o[W)Z8|PdŽ /]|ʏdbKJ}xP t$rG ot{ok 6JKHY] ‰M9gnke$p{xqo21a쳢}=镵\7k7%*DxM߼.Mѳ\v0!vd`y:.v*?(ERC Z 1$m j{ѱwPvBӑZC7 mzIya* Ocǽ?& /x&2˯]VA2Q:Τl$E]UFn(JvqRm__I0ﴌ/ėc |I)M_FdJI]N*37S`NL*婀V`܀A@3Z[k(T DGPJ]jჵJu[%_hL0Z0i?9 (Ad @x-Ay<lb3%׸ܢxHxO1W'ZjyCtk=}js/·{ݔݔ${6@!b#sj TJ sF8HT925ePrFRpB D@H *L q F* VF2:O~E3IHoK,jWv*PT>)*S)U[O'Pݥ0Kz{W\.YrYKa8TitFKJUDvTg?^VpZŌ8f7'DŠԑ7TCGe7TN&QHTof}a3M 4~6}w]fR#&NS/%L d-1Rg,B@JK45 Qx^lM&"Rw7- ,ƒ\0y.lmPʿ)4=*aw(6W_Qmto) %EۼRoQ )__N x;I1S@I*soJszWTJ*ʩ^@>u:k;9n/zzv_,1BTo'e!~T2mt;iC):;-l-A1c"$cC0ݖqGwWF%1 _[RH}iII9BNwީN8]✕e2 r,+8g"p3DpCT%M{!9 ~3Gm|& R sU V`([Riɛ*MrzsЛ 7炝v֛3T6]_.P7 ezBYNs15 !cf=C*tm5Ri*l0JMڻ8I K!Rh=+JFL()C1K]lLОvǸc8 &3u"cQ$Pq/cDQ0<,CS68GOxڏ`R)"_鏛R ^PKO??׊'M'=WOK뮷hvAI:Jktof3{ߨO"M4.dO؇뻊ck#FUt8 d]`19 .bN'?&i~Xv}*Y(H?)mdU-N_iRnUk~7 W]7FGf1*B1V8x7ҊFr4SE+%Y~TMsJڡ7MQ$bR {_TcSmN0p(MZlfCXE!kzkԉl);ސy4JB܈!2`.\Qd.*Vrau!D AE^?^ٍT |3nX盃MdjZc&S 6b rxŌsʼTYBCt&xH;1K7[&(4ME-ˁq'·vD:t^dB'z*T*DzGZ s+{ l@pD{ǛlӽCm6AS%NleWpgՋ7dk@ Nv*A⍶I E'G/7>m=vը bJze_;_^@."y &ſJz:ۉO֟XlFC t{,- Bew%JXKB<)GB8?[&':;U45*8T"T r4~l7:7?{S12g}@SE-i'8#k~b!WB;>gK)>PH_:aFd%/-l׋q K'EEA'weDh?j,b!Xs54x% ZEU[͝Vp"o dWJr.Nk菓Oi2.OJ0칻K'<džsGizrm$6s39+#$ڃBwqZcA3)J٠Ɲ[TGt hήCQr*Шd~>EVbg$3:*cl.51lZ oR.'B|ML o  _۟QƤfOP㕚RoqҮXztRhj'G=:bƀI WC1VI:k(96ؐpTg]ψ}®؍ugYHic~z#6NceHr!Oc4%(𞸒F; j1uFoY l!:fp F6dOЗTڒthnp&E i2\'CEaye<53=`J6j*.@Fb1 A&RP7Ki#x}âc1~LcEsJUuRDLr1TZs L,#-Ft҇4KH?aBEAp !3JZ@3(ͤ|se2QrPVZS͝ѐ=Ξߨ@+9DVpsXMzuqDqR9^U+~2h~WVۇyA:VPW_z{Kh˥)B&^'/Av|\_q)j߉$vyǫmL UL;U)L%Y 8(I-;k9GpE,7Www?W\).bV*L"Z;nzUH,u3XQ/(J }N.~sf)%+x7 V20 S'i2('c5cUקThZ0,EOC.CMZ"NيY# *$ZCcŌıK°B4܋hߣ*tb Jrd*DZ J6-"jk,W!b=Bnܭ] 9u|r~-W(BOzz=" WYib]h0rvRu\gngJ1ƲeG<ưH-54<DPv/.vI7WzQ[;v񇙯AM`9B5hEf&UÛ]lGEOH!߰ =b\6.hm)!J;8 !n 6{rLpeLS ) ɬf?7(P9{T)xŰ+k@3XٶKsDS_d4z0P<1H$Vv4O4w;㒙F߿ iTH@qEVQd'|9CoG)<5׫|kܞ6>E+N sE< 1=m<͢͠Ƿ*47S9d^gx$TgJmFfu2T~)m3!$C'qW$(6pOFq\mEC'џ&k~6b>ò]VPd/{lɇ=q_?1>YC7z -NAxw`]Zo/^x g3m,0IDžV+:m'* N!n~7E\KJps7n\>EڌPqu~yvBw;~SAϯK<+S+6}S\,~ꝉhpYc4SzR8]8p<[4DpVWX]FHQ.et~^mc+pf AqRG-@0fV&s&J*6Q]2 q+wNmԹyq*BF"tne8U[+uT@)Fjf In vZ,+"X7:͚t4cQbKjJ7 8"`xh͘+$jwi#˿B d>bYMA'NcRx8VyͣyMR8`]]]> Xp[?Ąc-Gi@-<D2֣phsLoǻX֚OXg =]J^ -[qx٘s2^('*@_&.s呎1MZYI*mC&[)'bjpE a8J VJpRULp*WFh Ta-ȃP\juS_k?upǁdk \"?Sv)ϯcpΟη,/q4!51d:jHXn@1ֿ_e-;UK"RZ^2RP+K4~;R\P ęx]pߵ6tL>2UsFɟUz8# &>}2HA +wʲ?vv~c(fc8*WޗUBz w30qz\m;=]KAVwuxzvrdo}֮#ߠOLllߠ*G9?] m'_d}nǣ7Y޻cX uhoϚ'p'!zǫ^/tԢy7TQj. .j j?ӳ֛7-.w MK7 V}^{Uo׉K7nT.ݎTDkT@/vz۟ڡOBlhf)49۽ Wgi[4f`MUR= yieݩNV=܌_(ѯVT#pWEUkiQi\K4*'N1'$LzBRhjB-h9,!*Ѝ(,y Z#ai¨BTީpZaaFg%p$k@! 0xքidPBFANWx1wv\J EvgW 7_~*hwy 4aw$WP{r﮺' Frn=O٣Nɇtrk4~z ue]{{={_Fw쟣_`2{0ӯQR2zHT_^}kq sy1o1E-޿azs.d4yTUh-25~vOQyԉZt-sbl)U#l !Qݿ$ QLcJKB <<TD08(8}[Eϐ;<) }lƘ g*5Oʓ>|xkg)_㼼!µK}`6#<9o;uIu2SIO#}uw9@e[cNO膃}5BɰۅX˺KT` ĜNI!Bs"}pOMj .Ȯe2$u43>*x`LA `39Sq$BřZO/LXBKyLLJɗpD6gq$;0-8SpcG‚k*V@")}iReI!&zfL"uO`ALR"6QgZ*SzsO1U1s04]ɘӇgLĉV %|)eZ2ҋJ]fE|V2ĥ\DTr2D2KdXkc8-Y:D`XG?E`["l H|zi!Yh2S8.߼_ P k^@ºAv'o/ >3$Tcrj]`)%;DA"+Yp1,D1.Niz#n6RkKY?jCQIG ,"R"˹&rFaDDDҹ~)6r!!wYZ%EJs5ɀ5f]?Eֵ9,ZL*4AK!X*܉s%.?wqI|ʛ:T$\/Gͧ]¬(~R=̱}E-k>D3 jhaG(&@ӑan˫O ;3K FK'f7߷A ,mZZ2T|U?ow.‡B~Z~ +qX0Xdr߇^l86o3䇇`@6IKF߷(i8TwuUW{ʽKJ/K[:Dר(V =v+0Bdn*40Ffͬ$\[^h^ALYPJk. 6ybߏIVQx2.vWN/dZPi7sU_P6 Iݎj+x~j:t~04a A7}1xvP ϾP(֬9bm"~|j QP2Ub8`OV ߲&eS~s>;)(Z3o 7.n~}6Bq!HE`WgM`6ܱ`\Ռu{ZDp]qay&i?i=.Nh_ݫiG@)8PHahRw cM[?+^+dB LR{jJʢ+b,yaxLfs*eN/l l/?@F?q7V#oFX)kNL)  ˜3)@,$T(Baq(dZr:ּ-;pwVX.àĝ TN>HJ~0\vHGU6ƪ Ty3a)7B\ύ ){}d;aS|ye3n@k`^fdhul1h [F1ʏ-eIT*l6Ԃ(@0 XEARAD*4Ls0&\*.ZXSFXI$YbxgHQI'jޟ2T)3\{x[ {B)DRj{ 2k 4=ggUoK{PN3 ӔkB~< #pL<1̙|L.lsK5)vK\"=6Z:*%բQ㶁a4 ,7,Hn3NGyNTJňiF2(6S0WC#c$f$ %SDQji*OzUDRW wӕ(I'ʋ0o<k =52Kj̽ W/f IIgy%-}.yXeȌUvQ"ZbtUH .+Ir~>A4#懄!AH5` n  waiե:88I7"GNzZRƚ@s <7GD^Z ]AN;;b{Dq-9ecUϮhro=kD տcUƪMYkGD$aK$ J5\ b2,$ OEyߺjEV~ ٮDvgo"kRz DplXLQTP*$10"b͝AUn.#K7WU ,^UM),An֋mhf_|sxW_j05J(B~7l SmZ ܶE28 MLDhVG # !U"Xr>H+:W\ujWn&&=/$ , EY,BF ,NicD)H# \F,5\RjBi*> o?(iBeh"b:L"`rF(4$dJBbBJ"JRAA$6}춞PaI`^Ņ)܏pq:r &#*v. 0,N[LFlo򐆚^'TQt2 Z'K[h]M5YOQvĴh SbW,SxbT8AU3w65S9 `5\P&ddo\25mS:i,j;o|h0|RB đs9]K\BJZukFVH:3Q$Ub-c͒:Gb8Mn<ӴPϩ0D(Q{(1l/I>%Sh“{w#h3T8Q==ZrrvWNbf: |! f˜7ͪVlJPÐex۩Mm $#Nt…[{XJ]:Zd.1XcYi^tUɓ($1%f)9ZrpC(\6|")L# nѧQ55qM +I>ҿMpTC6gq ~XFDlq&|vD*E& 2aܨ_gj<ߵ@åޯ95x )^>:h냻0ݪ7I3 ji A<dP~~ u\$x cA`nfܞG1.%đre0QeD~ӄ 8%{YoH8amC<Ϡm˄=\&|mgi20?$\}P6[&ZIlck_l 40m4y824_zǯ3֌;ۓmP/VŇC9[4Z:Op!E0jڊp(>~i&dÔ_+ݨMEqD7+ҵӥZp=ﷷ11˨uR|Z!DCO|4ӍgaĘl&+opQ5b;H#`CbJ0Sx ϗm SQ'5:  oB7qL(!$B!#ðH @)eH (CIYś{tҌrW׉Ф'-_Aup1v@:$VJ5"Jil,w^1,& S-&Lc[ip̺EYcP;Y:sߤ)ȟfȂq2ךR%J#fIU8Ǽ4ԐHB%dFdHHE&`)J-U I ' :j}WKaKNZ؍$gtω P.x3kgom DJ2R@iq2,ކ(J%;pLKypr4"rpJ-U{ g,OR1rv7V]oXu)k9lJRƹHӔ ͣ8"&HF!cB1&*MJ#c)gVT)lj)5庹žizsYڕ=qIm@=s'-,nͶ՛Uòwz@Fk/ &ɠWV)}~wOYx/2.];lBeMֹ22LEK~: e!D{&qcKFA=fR]bpؚBa~%;ɃkD9LoY/5X?%.t 7t36ԇ|:}:g![Y(Oq^S8yтSIi9x7pU P(%$'GBGeCHAy Q>`ɰ @*0&3BR(QHbG"=@]A%FXrs aD#gz$ |E1𰏢8Fj$9ӧh ñ)"A7gb0!%4J>i}^e9Vq`:g>EG]ߡ=$0շc@Pߛh0rq " S}h|=agH=JkTꜶ'ĥ,ͱ3n?'FXCPn.!-ЙxәnC'HzY2.^8y3wE\ŖK=ZBSʷV,HW,S(%;⛝>8DYcRbY' 6An6whF5o#|Gv5ODe߿O3l t @?czS5|ߎN],NB-ΘϻΘuhi.>\YÕ]0j K@ޯI- >[ ]dWih/ⴃ xȵWg O7YG ,Py0o'Q k)nwy5mڱό~y0Jك '}01Z@ϔQC̭q˂C.aBef­p˂VZ!f!a7C&Ex^?ԷwCvBI5\x8Ns7Nsp CJuO3q斨ܰ%:AbԵ<.jM% qs]HSܺƸeG ͑b7ǺBuߗaw58Nq`Ja՗:Dqg ~Kgu\R+LDcz|6Fi޷-3GMS@4=gKGS>NW=ZʼnzpjvUHXuJh' l&مԲZi hM𴲑*4-oV@z4:fYA-l 'T G_,9({cS7v2LV R!GVݏ۹7S}enGm g8V[|wݧ66i_S*f 1 : ʅzH(ȋ%W]S, $42 Udj&eb(3JIM]ʅi H#"PAT=mAHB0ꙑGak pcfPj RT*Y͖"' Oē$ tR!,xR7ٕ0tF \OgN~!t$\0}Gj.YrIEyg4M 5.- uJ wJ\+;jL2_o {*HbS\s4Q6j?VSœS⟇ط?UhO &?QZ,! *ž lSS( QLyM-(&IC BU78iQƐVt-%IWx F)sQ)M' |[\u{J7~BEzO  {:>dcG}FH˺0E>; dW岯V!dMH6vѱ d1Ir Л UXz+ŊOy;=x<]~0{ N )PjLO80?:% y~9)|6?#)e<h3D1>0w0Ns0DU:70 ڮ{!ٗ`uCq=>92R8%b5a=F8L&ʍn0xlv`A=$ìi˶{ŰTaeni(RA-",z|۷UKo'[ATK.)wtrDWES A~D4[ۦBb :,^o)w`8i: M"S"ٟDK:Oh:8tp!8G3u}rACi> &.>ؤwxVvcDa+MrN\SرFi9 %.(Ed 9* 6M52rq;i Z.LkPTH!g[%Hí#h#ca}ծ}_9e'P4-ZJBQN!Y'4H 45ҵ72ҵoWO`] e&~µ:N\8ͨ+qAqF)4էBi9pm;q٦p޺֒q.VɄWےb PjGAi>*o h9ц ޅ (l]H2T&8ڢJV!Ŵ]XDB ҅ǒ'C+X >~MBF@*hbpL|Ѹm$'@*T"DB >Bq"(72p1խV[}i5k߷(gA-5wk&"a[ëh:E`eF {)VR*kRR*4أD j?ӈ0M"Zt՟BNףP ]%+U*wM%E53T1^Rٵfr91vCuYԊ*caP |nmYzvo dV |Lп&ǭ>yY_ 𺬟}eKF@R|}ohF̙6lh8 1C s"ixC #}$ +pI"/%Kjֻj(o0گ(eX=Fx qϓ<n6o{;5ReMW6EYc"dϊ>ebWxany^V?#G~bfKFvLCG?t_TVRQ?:H G %D1$D Nw3N)1RS¹UstKeW..rq)to{]˹2ܽRʽe7>v0Ss?ݵ7kN 4_7=/XyHlTk'=mH:Tp~Ѿp]9^95P)Ԙ &{ŅqJ>)%Qǒb?P+A㲅9qeԑ5H5|NOa@|k-ƀsCČp'AnocK&w$;j,_H*KWQsԒi,wZ7Gͬ%I`:C z[8fVtbjN7E35yU*ZLV9bs7nn)ްPP li=qyW;._efNQo&nqS;-a :?7;G<--;pֵZp}_j}8ũv|ϾT X}C+k&ĎϞps#N;簓{ӱth)\mmlG!ǀ_7x- * 1FLu`7eqjvnjUiҶUF. O+VhZ 5U( 1@jSop|dV'-Z-aTf畺 ͯ-Pcd/nfbtJ2@6߀꬘>H#+MXG4Zlv$$ $9ᢧxTa \ 9* 8"9UhZUKhAN;42.dΗt@`PK$9 j-o&kPݏu{Vs*B]fjsr #er^`0s;Yg =j/KRU0 a~x0;|2әyU?u M.#\,zټC5QUUMTneu.N)%N_ULuS-6m3W$ƼIkOZ7sTWO WuJyZ0Z Djp0b54.:͒{=r/й~VhosU Y2{9&0%;X*^d뱑R!D (ZO ͔iؔ#]'s=oIq}=]xm̾β]RP,E+mHAђr6}U=pa;3+!aoi^Ev* 'RX)RQNdOۨqՂϝԪye#Yw%DTaajI<jģw׭ v[HNV@HcjEr0 *(B)KP`,Ipc`$mM5QTw Be?x⚠|D4J5[H,g(wnkQy,U3BO09Q<Ҋ\BPcaK ׌呏t9C5D1|OЈFDS8B>(_e `/vyI@{u d HZ>ǁrϟ6UĒ_M4_{^Iar<$)npӏW ?\}wkHW\wҽ.ޕE½vuwŝx]|;O6k@_ONl>^~D圈lH؍:xN ϴ߶r\ZҟƓNr7x{A4nj+JԦ O.a:^g _Rν,Y8'ü4׹/(sO.6XXWv8 G(VrcWŽJ W&x/wxIFs3+]>u -,Yt;OsNNxkoZJ]no/ZXuqM $vT.j(rz_>f<":jjeV TJ尭d.JU2[M"]m_GN'Y~qylt'Y)"a'lճ}a.nԲĴEQܑRtEY#:g㢷UvTx KY\:OVOEDq[ kF^LrJ7.w\uU(Ƿb)EZb[_ц7/!]2d'\ ZѨ Nenl_ەBж!+m,}\Q)pd:r{Q=L[ppnD}o\" _5 Je=oMX 3Nm^ltq*K%H'O8ͧV2`t9|:Y<'Ɋ V*n4*KN˒֒RGXvi.} sDN! >UhogBh(e07^nOk^$J3_~p/2^4•X4[=e\o}/H]_0$SBu"Uj;~4p’΃pO5%@r2rt>>Xsr}~}S{5u|g```GeqG6&oT}FJkn΢EXŲ32ի 3ti1ѱvb-WsSa]!;lT`rT_=[tSɈC@H_=3|S6Z :1GY?ۿm-s-q,^ձA"($򌜮c򷱐 IJ2٘d_&ŒhRhl2nĮook-)dy5zOw&*abN̯%sNJT=|;v4&S[n~*[9Sp;A$[ ȞAArv-p0l?c; I*ǿ1 D y56Irrn|ؒu9Pd4oWz)va2IAhBiX|] \Fk0d{Dkʕ5*>| I82!:4玢 ϋǞs9xWe<w*G\%qyeׇWUA# },U +kDQ3Rʮ gvTÇYu4Z!V]>g42s{DJ#^ՀԊ".$E].}S7FBJ/CjaEKUp(F2+{ua$y.̊wQGdwQkħ=䍁TۯwjD#;\m @( NY!H5\HgwoK8p 0qEB:  BI5XDDiE }_/1A ˤCƞˀTz # LA)XT>H+_ː{Q.˻]^R 5TiO"#jC 9GAJ?TK |}J7|9z%dBMI֔?x{swvtŵ˱{'k])]$ܻkQ{Wܙo{dSTuzd*}MrND5L.jP1MsDJv V m+ǥ(.i<.J+yDx̸DmUxv %EZ*dW源x5Gy{ppźRÑ8nď|.o3[~a ܿN2]0Loa뷓4{ֻry,o~LgʮBir? ֈ4y4-z\n/<[uqㆎZZ|N Frpտ9ڀpmކosQ;m l[^Ci#㴹"-Wt]576J[@DUزi1/s WXcxkG:SfOu8dfVGM 7(x侰V :ez 餿kO{QѼ pjJ[^qXS쑨&4k%݄E` 4Uj=RVS3VYAEMU,5͎vgK aULsSEfPd++mؗqяQ󖫞f?RY,ue͖!Dc[(nd WjET#@u 5ۥAH[STwxZݎ,1!QRf!bUxXq FC x͋V}+E_W ;߃Ʒc' 2Xcdǵ 2Rm"&7Mg*T̊\'σ%%5AOrh\ncpa0 WCJ ;i7_fѪW"x!(bzZ<3 y>z˗ p>b4Ha A4B>=I)Ϛb5Yy5jZ>W53V O׿x.GD xQ1[-y DV`\zMem *xm!3uƠ""1˩z =pğ2UX< \l&D kZñApmx\.KL؅PH]>FdhpMy{1P9đpf¯W,D:$ ނf+"!4q vk''R3£W"r'&4Y,s3 Tj\v_28Kl|4wHx}?MwI'L:_M'!; (ɉ7WSi$9MVҠ)ܼf0_tx2j@l5(HlHۅqpYN0i10wP i n ؋jm[$j9]sfytsŤr` tWUd\2OvPF$'%AHU-dkMQ+;T;^L#exB\sqH9{* GltW#<>οÛ #8_DXZF LQ,0 #"\bhSr^r TֽH2E%BL9*ŨMN3ԔjX1 @B 1C"f! $'&DB/F)%" c=hGS$#-T'5>4b2(b{~$>Ҙ [`"Ri_3\ |x sɐaQ!QM9U(n2\=BSu{&ORo~ђc+|'ϿmntEӗ PϸV,]cGGO#k|uANִ3)n.xXxJz8(iFRӫJM~\^ &bIq=~;‡)+s@l'֡LL6y(*vo{J=GtyWE,7m\ќ~*iUI?0I&ăw6,f#U:vHcfC1a:#-Ib qyR英F,&w&/IhU#x%#SG\q8{A,}O]{ /*\YЭ|gd8^M+g7ʾp|)|gOG2"`PGӀ:iq)ǣ عVP #@{8;: qԟ#Ic(vD–P#w:s Y;(} |/ /kej/pg|#-C B# ‚|GH7\"*x`X.<.uL@GHqJbǪpYd"Ol޲@9~28Wٻ8r$W=`e`p?4bӍy^xʲuY%}%NI%;KUT_D0Q͆*$C+RY^nU!_f0+88xUC`EՄdψٖ(l#( !۞j{5X5+TU DMUpQVvH1\1M`|h{rjQ@10Id؄8E&J./ǶYr6іӢ&RvBR#S$k]ؑEPecZؽ8/md_Pw8%olT (VP<+Kf=iupĊRn9-XZ.M+%LimA% |>sxO̙]- ?E/[ct,EPZjChiܖb30euX:i3<+ aiVrc% !缲ϲJB<3X95J&Nn!lZJ;Z:ZbVk*Izz14Q$+8@(rVnpw:"lګU.PF:a^Fk)^6V%֊^X c$;}`AR-ױv|Y!o<ޱdI]Q8,sPXֺcIӚ~]r&b lHhք'+O]/#z^vtYF26HQZ:><i!1, CU9vNIE)\j]"fK=\u}{6M+m?J CPIE--n%PGa͞ v$kNt.48he'qy]@W-C6辰6fB_;Ŋipdڧ"lC̈́:mMm݄t#Ua,ӫ #z5!*J2R?@).|ѿ-\oªu~-%jrn\I%O0h$eͪW-=猀_Q^t7f[h "1ѹP$ 5' 9knyQtZwgj*xpC·VizCx ]}nZ nvk.;EՓ7EK9=s/yYwxdnMȑ6p/`?K֛V\?._R+W0'vك{(sy9,ѻQȻ%~߭=\ \^)u ${GFFC̮(VItFXeab2mr_Iu&e1v!3bPk '&ZOHC-SRKk5@z֬mYۉTVi7bI9v#$5 F,I"ލX1R-ߍX2/rO/ [uU^7nG>hvhX&+ضm~qO@(/g`ivVcڀ<358?,;W机|ݹX=RߏߟvoW;ETfvN J~[qZ)P*@; @z{T~Zd.n"A5Ԓd.C5㡕?+ٺ_JLYs 9霓j K2tГ0ң? BĔײ,*VYej9 i|eilbP(ZV а[h)hYf{o?x, oh1-U\s-*#ԖYcүnUIŇ,yƵx/;n`| \x5vڂ^7TrA&nb΂Xv٢3^k{ˑݶ{Mi5vgϖqIސ]5c =Iu=œ^ ~׀QVwꊭ$V/PLo=HR[=='NG!O3rAJHWhwIIEK#%r ~#Y~E,Bs IJٗ* *ML2]/!_SLAVo'%@(c ?Foe(~:1!2S 1N Wh}#FQ4W)] ɤ~ 8L&L;Ӵw]i_eZەOKAT̉Ͽr=K5AKҵ2__~&|+e-O[)ֲRJtʈZtvvŎ_P}5֑uT_s܄_Ppw%_R|c8;!yk8-#1ww?GGu6OPKف+٘vјD؞wڬf+c]W+Ϗe7g=};[j섭5ă1 F*^͍[oC.5#:06[A@F2R 7#W&m1bG̣>9\s] K/F,G%=A# qN;3j/u{zZiT ̠>̼y:X=T]m>y:2y)/Za(\j@uZvbkAV{ܐy4RKMJmV4%w#N?=ݑ_4Y"y[g')6C[{ӵnB = 3.0<*ۻ?{ȍٍd/!L6;I&p&0y+K>"jn,7JĶZMWb,~%~>0.xݜS 0H~ChQ(ȶ.Brog 8{_h; U?"w ߲=-AhZ[u8;tZ<܍4JF3)f'B_n<6EF.l6LݒOLszk| ?_|N$i:P p8@]>;FRr9^dD!{c Э"I_݊?{˺E(=PܗqaB9q>]Gs;jz]J(EwY!:8Ȉq&-4RZJ ܩXk< . krVd♞T40HV$g{-]:=,[w>o]wVPNC2u^cLn׌U|\-v>lNR%scf>s1Sܻzr}Wn1Pw5[MǗ~Х s2/v~z[L,o6{,rbٛT=~{0]\.-Y;Oz\XOaJ>1˒*3,䕛hMqz?Gy7JW;nNj1ޭ[:nwB^&T__#}FJu?:nNj1ޭU3DҚwK감Wn ~3b6?|uG 3^l:rу4ƃmmq5~`80_]`~ߞ~{oO_x|4L=5x\ۂΖ`<:g>dQ_ߵK7fJ tHr\;aYnҖP[ $U.*aL*s!,\\:*Ajo3F o/iOHbryΞA@]BEkg0N#g|gR eHhL>0 &=#O\G$sʂIϜd{dntrGz&=Nw6buX[|ll6:5p-/'r(.7mWuhx3 @$HLȝV"">? frbэ>Byq·l̠˝@tio𑹌1 p|}} ӭnu \bS;zu1dú. ƀܢ4sJ-ZH9 üXI${ U#-w/zk/EbS*r TM.sRR: _N!h&JfUQ$dG໽^~btFT^Jw^鴴x^JrRrM)zϻ`[*!Fw0bUk-^wB^&ٔཏ8-%vR116b[F jͻnuX+76G¶o y&' ן@e*&ٓ ODv dOn@*FؓC%!=vP HdD8!2,lDcl6lOHOb2Sݏg`GSP{DB$N1GPaC@.YrV!CBs=q(Z$録堙Ꞙ3RA#%%-(WvYӭkp7jɂ(~{6RIr[b#YN/Ofnk=]fӲ?.㛡Μ'7˯?5fw~g_Fo' قn|sW7>_|Y../!#WnjEzM}nhF-~^΍]-_pfL{|^a-t;7=Sy4M͙mFCB8sePK jyُf, /4sцjuAu-l2r _@7 >cAwk*W&! oivsokNgf0Rh i9ˡY8ˑ5@ 8 #\ B` 6cr_F7+a7YMwn2 {5I:۹}xw8cۺ LسZr"Q[o$p8(^k S䩜tx|GQb0S2&pfMkU(-F`$@c)Mj-ej5*u.H.Ɯj jMF03X!bJL&B"n)`dR÷! J!S#N֨ojbIR&& le"kP:ѭN* [# Z&?KsIi,5!)[szÇO(6[,7u>{G"쫄AkPIX&UR]M%-p@%^q6I%JJ$|BLˇ&I*)mqTa_% v?ԩV*5\RJp^KIJ!!ިaJeUH)zF 7|=jw9#Dy^l-kSr@3YݣEc|4$S*XaP&_K` _)0~W|w+_З㛟"q q../*SX*b9Z3_Wٛ?cIJ,#eG6ާo?}m!(L=H󛙏7:,.GUTՉrj ! `Hr|9 ` Gڸ9f>'5}N܄$c=M.crX~JayW[eqmV<|݅in/\sXp$ONFPϧΣ.a y(թ&P:ͩMv+[s:姻cQUz$Af<1բp^=9{%8_@M (ABdoXNVƵ VkC$4JjJZJcE B!pLb< 10A)AJL$`z9r"(7X2Tc S02 &0'RUNZ* 0([4^BWo  qsasv1oⶃ>+n&Dk!TSq>ﭧ6laՆ;l[? ).8aDE?]xuMF-=hj]}?ZuF1@N Ý Nͥr6CFR #HĜ͑QrN P{KZ<-.~Iμ((Q 3H̋R// $e)@|U@ޚ2x| v3xQ>|tQ#SHPzw9RU8*8b\etNDC@c)Q+rJ&5ap$ ĔD.šͅ`X}FFB2[`aMƂmSqO1v~Dr덱eE%El3xP:|OuQxxVJX9|G 3x/ BDd=EI\.Jθy%8EA B_]h$zy(.cEKG`yFtƴrs\ͬv4F*#s12jd 0TAmM%凞{ n$9U̗c,T^Ou}YLa\d7+}\&ڮ׽ -$u.%D4|!7l$u[9T[~yL璅2!\ BnP8Ki J1JjBZ8xy Sf)XO/ye iEK kdJ^A)YJx )VsRAS\Oc/&:h=@$ j(`@}\@ud5r,$dHC1dXA$c:SN:#+J[*3HPg ge2Mf6F2?_ʁN;bNT R~=˱F\)BP!3#DPH[ '3UlXzfQKnj:>;7w! 6O] nR.l5Y]1(*<||MMI/iA(bAʈE̐pDPdSIg%@Yk-an &Niܲj5 CbonDYisr>*Du0R9>n'gfF*0 2(Rc51JqjȻOǩS )yAau.FAnesjp<|*A CՅ88 9p o17Jq05?{> V*hPe\!@)G{=2P9B-)H{=O;ǹ| 2ₐ٩7*|`~kz)Fa;s4i^{Ӽy]6 |<Xi R-Ąd \Tn@ *had\^_*\ No_CJmjP\!@\\9fݚ3?-m_ݧNn5s |)^d\ܰ "Zrˉ(9Xq P0|bM-rv'VR#CKj ýԱrh.Ib03@ʱUԵ @FYg) d* RJm:uаܧrs;̿:*bƪoqo_ 陈ߡwo?Ȅ(07~ۺwo.@lజ| ݌ȇυ_>~{p/5?"U@!LǍ}rDs#OLȽoj:u~BG1g:by{x(;"a 2|!V 9ʑBpAH2g1H:W`$3Ƶr;! dF&ɂL ({If1D΃P )h+݌SC `{V0.Z,F Mǁ=4nrJ'n^"S0P:bvpvqf)!8b{!F)kY?{6J/sH*?>SS٤I^6S.cȒVd/@R2DQ'USF/p(5+f⑰wC͢ R=I eRS*dbWxo9pc2d]q UL30F"'hg 'd4{ċ뻚"g0'F'P.Hu|aNHb9>Ht|O ݭ@(uO;.vlRUz؋́1وx9U;O 4c m8hB{*(ASU{ PZ  Դ^!OC@0y:f)~gȧ˕,ZY֯ qF3eiXuwI>U>6I~5H3^5H~zț\-u)?]i{w)OG7n2R^ Qd%ضS_{̏k<ͯHDWZSgӈT+Hy&4nApi/%1XG1` t.9j =IZ霼ө68sxېB {ݪ7ND97~Ťl՞3)P{H~9سU[MnW_x^ey[WٻO`J.{v0J WE.JpCؓبO 8[}i.-)$K8 y&Fdhil-I[3cu$FW<kl#\oٌj׷ ,7:5;,}~[x/ܜZ}>8ɨw_Q\o;_I}s(sWŘӧA2ot5Xjc!NU',[av9/-!.Na* Ћ&}dUYDj Dhž{? {|38ٓ6wI8tbxB#c/vio(;. ݣze/fVɶWsݛ^ a]_]AćׂKq|_pqoz#Ę (xD 7e̝'S&Kgv޶ b56Hd[ ;nHd_ @t_SO,h"n!N'D%7h)Sn]EfOrcmqLNVkӑ3K3d=OI5Q,c6 i?'lU6uvTFygvϽŗlf~oJ~'%s60}'n5&f^y G@);$;h[NOV.:XFnp[uFظs.K90s~ayn*aetlŔUMjKp t.Itb@I!IPIKŇҜ!BB0xZy3lkÞU_? {M&_?>&g{f>בWfd bNϚ;Y¶=qض_tdŘ5aVŖݣq7pQ(*(;v&mCJhgljsGe(oӧr̐<3j.K396NPPT~-X37{e` FOkܚ/ȹJϮ0$~͢{}Wkx?!z-0 {&7RLzܼOcɬz5j֛gV沁=5r#@#5Fb:!ĚhMk0«W/W>OޝC9y6濓KHQ0"ҐQ4@ (!uX)(P9!0' jlY髣َ׽#ГD17~rZK>,+Zp)t0^>F\_FabVslH`©~s28u7M&Kӆvp |'gľӋxfYdM@Ă,HVV̿llzCxZ 7צ^@֔ +M)xi".ʰ@cvZj1.ӡqs_JNUx/^k,D\ gh8'(TPHB fX \IP51K8E@@b,TY(^W^;v%\` A_" BZƆq ٠c\804 !f!eέ{6j{&O-22WxyTI >ˑ\FPĥ8`ڻıD`jbi3d7d9댖..ӐTb˸=#[-\X2yh%e}^Xsn)jʲHedgsYt"tIy=*28- U1iJ;\ۇ^$26bٚmbJZ.ef?UtjnQ$@n_ <0w{=?ezSV"ZJϴ,X+[IxNwn系Nڵ[J_V"ZJmk7ZGtʾv;`*ecڭ|zvBB~p):H%UyΪtzN!g:]E r|$5f2M'8q_EDR\c*ֆ<|TeS:bP6`*֕1Mw\i-n'0vey uLFp'1Ėx'1xɧ,%d}:96`sXVi$`唞wWD}F+Fl79B˖%xW78,FiXy"y\mOݦ6=<,٢ ^NT秇X)f_g"qQ"gGmY?OoڷFw\lYmlYu*~6.D7]#Œ}}G6/L@ۣ|iS]P/gCp,^xݛy&Oi>b:Oz,>ݍ~aCI_QYNky\&~ӌ ߑO7W[qy-:A~D nafEx$Ƕ 'A;h.ێ<(eخ_phl3Ohv?8?̵)h, Hm =X1fƹ^{%[1t`v6nw֦oau#ʹc|c$l@wmFkm|*h˂MCk bMߌMf~#k YY{Nǒ-16>Owc݋G"V ̒t:‘oKno]H{$@q?"ߙv֒t;!gs68}8"yC%>f|jFdFn2*폴6vxDhw9K>9 ls*{ V4lM{-܃C4hPojK0[MrqAVsP5t7dF̸]g4n0d!¤BGB)4:äTL1X!0Y̼LӟgYf4xCKZZn4у!a,E:[ m5 )cD^\K2Ou?zks_I ;H;F.m"Xk2y n;.mqrou '㇄ O7f =JזA7=|(Myw,PB{hf9OA#Psst? I>ZM,G^^t-Je3 @!r4L5$9(*H Nm,V:b Qni^%O+ԮaTPu ?ޛZ좑`hʼnEE#ZlRLIVz*ы \sԳ8s@/:kރ3(E^@\JYNY խq"<%5^!k:-2‡ɓX(Gs.16`]M:`GcW2LFbqqcĂn^t$M= t*01MtY_yu*JwTG )QYyzkv}(wFM}FhvH_d70vQXDXIE:W+dP*du*f l00ϡܳWô/y\lO8+HB6Ad!Ag{N.s*HRJYY7VZ8sHf%đsw&|Ub)0V_b?{-; oxa{ `ݡu#BIZ}x0keGJ5H]| w>W~< i5nnuwy2xqjsk7TV^|^24#]z|M}7d|?nڵAv~߾1MxbN%9YS<._ҍ_ lοmi@^ߖ__{ 2+GV>oyZEgwogbz,|c1oNOrfϙJS\b*[!/=}G=qeH![/-mcup̩!;̺uպNNYenx~G֋A uzۘcvDnmnk0SS-L}}ty)ś'i:2pkkP5_??=D /&Gr! `FoqA[*m 'BW', ^zXf R_xYPvG}ыAmcѪʜsC%' ylIy:1hJbƬ<[R;k@lN.N=⼲n`z1hNos۞^=^@[ yҜ%ݥ(naDKr)&z V6|Q˺C)FrJ)gLT::cɉ/(梄BĜ"|_j 2@{#,c%C]:z@5B/-ƥ9ߎ?0Meo pvqw+f1[/-mcuQȜ^ںuպNNW Zz1hNgێ^֭wmpp œ]T (: |YGF`*֕Q4  sT5*PLMKBfu)Z2ALc0HI&K#.!"Jpl[!O\`6S14(S0C8_NR'1WH 8(̤cK3_bO5LP;u*&*3 T&beHfH#17(%8(|t_Hd#NUAs)ƿz=QFiEǿ h)N1dݻQX#nĕAuʾcVct@1/hw!qԾqn7EY %߱SV,aˀuح[ StoNR8tq73tz׷ڳg6}o_k`in_/Ez|W_AŬ..rjmU>*Wb18Mj8mEҖ8%,eJەi U5$B MBm)1j.۪|H (/Qȋ1#N%mT;ԴXi8h]Pdm5ʫ r}mm*mK X-JyEPSD+V޸VP4LXLqLʇP򊢊Odp[ vlnc&ze iMߍf!#zI8.w/|:v? v]`<~ude,+66Φ `A?Bɋ{{%=ݏf>gkO9)6M,' ôǫ2F*w0ɼ\&So9ʾqUkiVWWM 5wNE6 nS"S a E1y/x*ly!;!:4e`TpcS)BɫV [\J*ư:n/wz.=~wJϳa5Eݬl^Kh҄ͭLp %i# a8ֱ{DCb0W%O.tz^ꈊbj_q@o܄ߏ\W9r4% 1{w^y{6ױ{MϣT/N1+~ IJ^VWgu1Bt[ןS7M{Xm)O {Ul(-={Ņ̋^֤s%sb7Kԡ{7k=Օ82%Gi& Hc7gnH\~NwEgim2a4у$W.:pMowm\|pӚَ{?O?ONsv4Z\_Bv,~]_ZO޿fs=]sh /ȩ$mҦLt8?\z?gLqdL_7xlJy!2;Sk@ƵԆw pA.Kz +KvX8.+,w7L^j<b#s1POE6N 6|Q49R짪pR1>f{F$TcC)L;4p3!B4z=tBV}"VkRi e)BRI  k%%L $cD(V3aMo> 2@dylZ26,Hv6|8E0"ұMK5p0DYH"4# 4# 7G>i3"̃,MBJo- dfHBhؙ`1wQq ,A4W)/HH;b.VZNm'UN#hCA̗%1z!:e`ϭk)M8f\ E,4JU*(1%N>AxFw}fsΟ+L{8񙧸\<%yaݯ2|+a ,ﻸ{3A!ߍmէރO0rNaF 'x2] r3r„QP=:z(bXJԊz2Qmivg#ԋ|y0Rfk+VԞG'$e@+xotaF,pRIE KǤbQ{บVF[xI@eK4uKHE=]nfKO\wdVBO:h_(EYzl^E׬NN;pV(w6$?@\mlm:@GXOt!`G]XTG|j,Qoo>6Y_{+] h>r:o6Bֶڻ&+ ·fS쾝PpfnMM.Gܥli9m_Pb'YljY3,,ϝD7D~%u,π}hD Q|~=zZOW ndnh 1[ȭ P~+^9BDc1%v~Ե!],7̵uLo}_7#̃58<8 θvD|ƑV2wW:wp YW+:̗ |rt$j EB;|.GXnW#s%T:L)% :>Y3Ex|\F3 % K!^ptGK勼R]1 &Ƕ۱b˟1}f ?-ǮLsXoɷhxK_Z'}ieK}ܷzr=i}eӍ?id;/eO*&g^%O! 0zG;d:xפKHCbNOKABYB$4'uC\( J5J\x'pG0Z30q]Dїg>dLІQOGۛ뷻%/? \W ujg;>'o )&][!@nG GTdsbH4:&<8J= RPdQL#&Alxi%P `E>! ~ i<ÓvDžigg@?rdgoՅ hV?ѐ;Bؔ6TwbZ' W +`xU' $1f?#BDZp,ONRA!4Ƥ"N@O 6$AΥ*?S#1!10+RH&H[]8cJ2a"8nFh5 ~iSr?6[g^s-8+9st駉M?ov.ػY|&& e+ G:ȐXapHRc1Eq q DXo.&ԩ^ڔT(lzRCNaDDXxR+41)5N*lTcb!P8*cCZ=<]_y8ߖq2/x:JfXw/dFe;WC(@҅|ﻸ{zJ!gywc}yWZS2 "\}ఢUBeͼTE|ӊzyV{wm 3b,^؇dW q|6Xtɵ`ݎ.qCv̈kv>X@W Y=&BHaJá FLw҆yoWS qvU+$M.'TƊۉL߸lZ ]vÓT9|լhE|&oW+mo֚\*=>xl|⇋oWwE8/]<'[XKSFt³v1YI+Z+䵓':샭YaYZAWEUogr2~9R95C3HқuЍnlo>7i|5SV! SJN~~tm& :mFM̻? ݺ@;h LA6F/p2 V1i6bm8wk?һuw0S'-ċk^1O6 yY\\>l\R@A'fӧa?!j't1VG9XoywnoGYӣZ.ɚ:(Q Z j=ce]SQ{gÇ$b:3eM[I!PQ=G8i{=9| @94[w5E*\>*T%L9tdp2gݩƘw8/ɏo.8;HΜ}(C[0E04&6(tlP;2*T[s.(d*tx(5m^l(WK&Gc#]Q(Gvd#[;:9GUD Nu4ː)eFLA:eJ(YU{mU31{設6hCƎS/@PsVX#HMPYtD,\1p8fPh\;}z1P-g#V[.fyX%Lylbѕ Pl'BhcpK€&$@đN}z C.#, |T0u7[*WN[(Y 3lH>MS$%ieJIҿ1h Cx!-z4r>0ObGc - GI?FwymbqE 1Z^ T0E@mC\'v#>iF1=Bv}jT`T#Or8%C˶TR8 T5TY poMX== =`ES-߆b/oiQ ӗ·T%1a~ &*GD*GhV@m&uQB12jp#@ҍRH TtLkyfzS~fz?\_WU_=kv"|P`z5—ÚKflc%}[LRcd_vI$$]c(!\OZAztN)>pĖ`T'OY:=䶠ˌYD3mv,9z{|A#,2Zgl # ~: Iog)+@&®41QZ4&AR0FcfH|ZUq^v}mjք)5vZBju-꽱4Mqqɺ4RYep 8~>et0Rspa㛋]j"FiEO_4{T2Z Ӄ;btĐ`Q@뼗~PH>k՜7v{"Q5뤮vts턑pjo[pynhK߇q(E7K ˖qk^7gE$.Q skr0w]aN<,1Utb;uU;G /]ZYjyExP1T,d>D, AU~eB߂%mTNK0>, *dBkE%H&)ګm䑾cl1vdpͧ81rVܷH&h{TY󻹰uT{j_z 9oػ%\Y }!?N:,&^d9jAl1K\6Y1R%E1r GTL "\[.|*vS(!UV?Zsn/yv.rrzb4u]3s$*vp>_nFTO-ZP]0LcT+Άd bϯ談/S5’<_r`jk|,2 "}Vlb"b -gҙDZz8li5ͅ gcb,oܒBU_L~LxOfbg6t vg6ʻ槛=_gDwM8{nlL)8{%97B#Đ@tHb)KЖrNsgɔZYBed0c J8: F) "u +I$mlㅳI3p ,ȉ4pI(^^tBbg*caH@Qvyآ+-5ָsηa _qt>n O_[t`d*{S$Ⱦ\ۀs7YS]QcWk;6>iu&ZN iͧ7zkZbOntOyf.F'=I4#\Szȓ7( -;4M9f$)VMVEPNwq~?=o5ێPY#W=hw,cd]M Qt8DW{jyY[YKY$N9 ;cްJF^;=LILpg- #!83;:'VJ G!P夜n_miO[f~޹!-=.ؓf?nr@s},?}9|u{s7yc!*tݑQ4$廭y\ZHA) *;^R찏aơpTچ('YCjFxLjRK(%IERj f#Jy/rB$P#}i%sv(ERr= )5ZCAixۛ@~ϚR3ds\z(,_9N>kJ vuF182P*JI}֔ >ǍRPu@iRPZJ GRy5<PLM>kJ oF14nW_UL2 ꯥbTšrOG$PO곦ԔyQ!lYICJK7f5J12.2.0#Rj^g,Qz|(ՑKr$SȞK}Bj)f_z(R(R0 Ci)5hzF14L^B`@iܙӑ4.{U‰@i\(VsQ<K_*)5|VQ$|[)~E곦ԒGRq( N!E곦ԀY#GiT|D@iTQjsQYJ9NI8R$st(()+Oq: Fu@~i\io= `VC2oM 6y^%T.h,L!d11҅}˝yƹBnG7%E( +6(0//34o.~s[<)n*ߠ|EسkN=ŁlOݺS5ӏ~ݝ 7Oh +׷o˳YQOh8ʏ&c@"k|{: y:J>ߟۻ׸JnRXe&Y3FMȲާ+vvmJ̻^3YnWËoacx1Rb ӯ~5ex >lqv{bmAu3P]o>D:X71~*Z5SFz|KI6Q>tIZPFOuzEugƶ`AZ']!p% Sy.qJXL:FTU."gK;U(r:d{Q{6hC9Qž0 w?nF_[p:H|J`pIs= nPP+&J'39/Fh R8&UD(0PF[N*ҵfm '?v tNKVO}7Ѧt9?s? u7Wqbq:)i#w4h-p8o%wK+Xb2|D@Z1Mn #24CB,<;L Ŕ0V#X )2`sQF5-(iT-@@DN;H3;[0ד >ϷKZt }\هU˽ӓpW]OӮm 78!28͢i̶ߤ]JX$B}7RI9\O&zJQGJ@YegkιdrK+MlxkmYmH -xS:(psC2>bqNcтRzoA=EēNZ3ym:dA 3'y|2Oy5>5d]{_ݜɖb prr[aA; }WzDѪ^떲!r.oyMvdi[81h[l6f-wqaһ8(…J[~(ҭsk`tUضH6jBVHZSf[vQ7 4i:+C엡2.CPnʐPL5ӆMhvWmhw։6[NyaA؆k \`pDe iM2L>!.E_ B-e T; \{r|<b#%)ifj+A⢔s!ThI%7%%W:UhQ QL#sW7gt,gwfcoLoX$vjdscArq\0ޫNhTqO dv_hk]󆯷P^Ps/_2d1]ݴc2gh[{HOK7J`>1Ș=ʵw:K{x׀[s?>'xH7s0SadЇ|zÞ!_c5i@Y"7&4!ĶI6l`hX ڨ]c4wa3ݒ1qdy(O-Gl1`G14Z)D@0"Db+'IcuӶV{5VGu…H%ZXBV,@mn&yRNA0wj3ݩ|w@ r!{T_W}dkOyyyE ?sw!/v٣ock3fJO'q38G 装\]<28eP&klriãjkN3-?iLUݕ5y:&m*xYÞ2کZ8:5XE^T){dt.o|F輣QD:k⫗u`Ep2bZt\(|jq3 ^찂* ɯzQ@J-X mk@hm4sYO_ZGhU*v[rk`lQp#%њܹ`eRbc˔13Pjگ_J Sm4xz*ԘB"KaVD7$(U&䊌]#*Oޕ[R) c_E H oY*0/ JǵJ9t6 #k8K#@-Q-FKU۲TFFڞ W% 4lUb74SSa6V*H9,RN$C*%&+,H&D3N%wE +XuP If3Y@@(o&* rR3+ 眉cjkkh]WSaxH3!r?H ˸%|^+^˧Gv2\QX2x/Q7֩ C.FV#Еݢ[{M| \zJF >!K@ؽrDO!f67e5TN:eLCEWBbyT~44bED %v$єr̬}A=ZbB 2K}JjL_j336Sqvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003423422715134220403017676 0ustar rootrootJan 21 17:56:04 crc systemd[1]: Starting Kubernetes Kubelet... Jan 21 17:56:04 crc restorecon[4702]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 17:56:05 crc restorecon[4702]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 17:56:05 crc restorecon[4702]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Jan 21 17:56:05 crc kubenswrapper[4808]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 21 17:56:05 crc kubenswrapper[4808]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 21 17:56:05 crc kubenswrapper[4808]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 21 17:56:05 crc kubenswrapper[4808]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 21 17:56:05 crc kubenswrapper[4808]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 21 17:56:05 crc kubenswrapper[4808]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.383237 4808 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386123 4808 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386138 4808 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386144 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386149 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386154 4808 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386160 4808 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386165 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386171 4808 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386176 4808 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386182 4808 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386188 4808 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386193 4808 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386197 4808 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386202 4808 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386207 4808 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386211 4808 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386216 4808 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386220 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386226 4808 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386232 4808 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386242 4808 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386269 4808 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386275 4808 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386280 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386284 4808 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386289 4808 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386293 4808 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386298 4808 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386302 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386306 4808 feature_gate.go:330] unrecognized feature gate: Example Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386310 4808 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386315 4808 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386322 4808 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386328 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386333 4808 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386339 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386345 4808 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386349 4808 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386356 4808 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386361 4808 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386366 4808 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386370 4808 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386376 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386381 4808 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386385 4808 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386389 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386394 4808 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386398 4808 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386403 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386407 4808 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386412 4808 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386416 4808 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386421 4808 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386425 4808 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386432 4808 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386437 4808 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386443 4808 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386449 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386454 4808 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386459 4808 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386463 4808 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386468 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386473 4808 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386478 4808 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386482 4808 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386486 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386490 4808 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386496 4808 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386501 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386505 4808 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.386512 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386608 4808 flags.go:64] FLAG: --address="0.0.0.0" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386621 4808 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386631 4808 flags.go:64] FLAG: --anonymous-auth="true" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386638 4808 flags.go:64] FLAG: --application-metrics-count-limit="100" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386645 4808 flags.go:64] FLAG: --authentication-token-webhook="false" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386650 4808 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386659 4808 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386667 4808 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386673 4808 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386678 4808 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386684 4808 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386690 4808 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386695 4808 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386701 4808 flags.go:64] FLAG: --cgroup-root="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386706 4808 flags.go:64] FLAG: --cgroups-per-qos="true" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386712 4808 flags.go:64] FLAG: --client-ca-file="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386717 4808 flags.go:64] FLAG: --cloud-config="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386723 4808 flags.go:64] FLAG: --cloud-provider="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386728 4808 flags.go:64] FLAG: --cluster-dns="[]" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386736 4808 flags.go:64] FLAG: --cluster-domain="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386741 4808 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386747 4808 flags.go:64] FLAG: --config-dir="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386753 4808 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386759 4808 flags.go:64] FLAG: --container-log-max-files="5" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386767 4808 flags.go:64] FLAG: --container-log-max-size="10Mi" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386773 4808 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386778 4808 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386784 4808 flags.go:64] FLAG: --containerd-namespace="k8s.io" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386789 4808 flags.go:64] FLAG: --contention-profiling="false" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386795 4808 flags.go:64] FLAG: --cpu-cfs-quota="true" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386801 4808 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386808 4808 flags.go:64] FLAG: --cpu-manager-policy="none" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386814 4808 flags.go:64] FLAG: --cpu-manager-policy-options="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386821 4808 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386826 4808 flags.go:64] FLAG: --enable-controller-attach-detach="true" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386832 4808 flags.go:64] FLAG: --enable-debugging-handlers="true" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386837 4808 flags.go:64] FLAG: --enable-load-reader="false" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386842 4808 flags.go:64] FLAG: --enable-server="true" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386848 4808 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386856 4808 flags.go:64] FLAG: --event-burst="100" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386861 4808 flags.go:64] FLAG: --event-qps="50" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386867 4808 flags.go:64] FLAG: --event-storage-age-limit="default=0" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386872 4808 flags.go:64] FLAG: --event-storage-event-limit="default=0" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386877 4808 flags.go:64] FLAG: --eviction-hard="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386884 4808 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386889 4808 flags.go:64] FLAG: --eviction-minimum-reclaim="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386895 4808 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386900 4808 flags.go:64] FLAG: --eviction-soft="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386906 4808 flags.go:64] FLAG: --eviction-soft-grace-period="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386912 4808 flags.go:64] FLAG: --exit-on-lock-contention="false" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386917 4808 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386923 4808 flags.go:64] FLAG: --experimental-mounter-path="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386928 4808 flags.go:64] FLAG: --fail-cgroupv1="false" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386933 4808 flags.go:64] FLAG: --fail-swap-on="true" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386938 4808 flags.go:64] FLAG: --feature-gates="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386945 4808 flags.go:64] FLAG: --file-check-frequency="20s" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386951 4808 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386956 4808 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386963 4808 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386968 4808 flags.go:64] FLAG: --healthz-port="10248" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386974 4808 flags.go:64] FLAG: --help="false" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386979 4808 flags.go:64] FLAG: --hostname-override="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.386993 4808 flags.go:64] FLAG: --housekeeping-interval="10s" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387002 4808 flags.go:64] FLAG: --http-check-frequency="20s" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387009 4808 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387015 4808 flags.go:64] FLAG: --image-credential-provider-config="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387020 4808 flags.go:64] FLAG: --image-gc-high-threshold="85" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387026 4808 flags.go:64] FLAG: --image-gc-low-threshold="80" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387031 4808 flags.go:64] FLAG: --image-service-endpoint="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387037 4808 flags.go:64] FLAG: --kernel-memcg-notification="false" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387042 4808 flags.go:64] FLAG: --kube-api-burst="100" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387047 4808 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387053 4808 flags.go:64] FLAG: --kube-api-qps="50" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387058 4808 flags.go:64] FLAG: --kube-reserved="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387064 4808 flags.go:64] FLAG: --kube-reserved-cgroup="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387069 4808 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387074 4808 flags.go:64] FLAG: --kubelet-cgroups="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387079 4808 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387085 4808 flags.go:64] FLAG: --lock-file="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387089 4808 flags.go:64] FLAG: --log-cadvisor-usage="false" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387095 4808 flags.go:64] FLAG: --log-flush-frequency="5s" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387100 4808 flags.go:64] FLAG: --log-json-info-buffer-size="0" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387109 4808 flags.go:64] FLAG: --log-json-split-stream="false" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387114 4808 flags.go:64] FLAG: --log-text-info-buffer-size="0" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387120 4808 flags.go:64] FLAG: --log-text-split-stream="false" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387125 4808 flags.go:64] FLAG: --logging-format="text" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387131 4808 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387137 4808 flags.go:64] FLAG: --make-iptables-util-chains="true" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387142 4808 flags.go:64] FLAG: --manifest-url="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387147 4808 flags.go:64] FLAG: --manifest-url-header="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387155 4808 flags.go:64] FLAG: --max-housekeeping-interval="15s" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387160 4808 flags.go:64] FLAG: --max-open-files="1000000" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387167 4808 flags.go:64] FLAG: --max-pods="110" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387173 4808 flags.go:64] FLAG: --maximum-dead-containers="-1" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387180 4808 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387186 4808 flags.go:64] FLAG: --memory-manager-policy="None" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387192 4808 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387197 4808 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387202 4808 flags.go:64] FLAG: --node-ip="192.168.126.11" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387208 4808 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387220 4808 flags.go:64] FLAG: --node-status-max-images="50" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387225 4808 flags.go:64] FLAG: --node-status-update-frequency="10s" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387231 4808 flags.go:64] FLAG: --oom-score-adj="-999" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387236 4808 flags.go:64] FLAG: --pod-cidr="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387258 4808 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387269 4808 flags.go:64] FLAG: --pod-manifest-path="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387275 4808 flags.go:64] FLAG: --pod-max-pids="-1" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387281 4808 flags.go:64] FLAG: --pods-per-core="0" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387287 4808 flags.go:64] FLAG: --port="10250" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387292 4808 flags.go:64] FLAG: --protect-kernel-defaults="false" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387297 4808 flags.go:64] FLAG: --provider-id="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387302 4808 flags.go:64] FLAG: --qos-reserved="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387308 4808 flags.go:64] FLAG: --read-only-port="10255" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387313 4808 flags.go:64] FLAG: --register-node="true" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387318 4808 flags.go:64] FLAG: --register-schedulable="true" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387323 4808 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387346 4808 flags.go:64] FLAG: --registry-burst="10" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387351 4808 flags.go:64] FLAG: --registry-qps="5" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387357 4808 flags.go:64] FLAG: --reserved-cpus="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387363 4808 flags.go:64] FLAG: --reserved-memory="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387374 4808 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387380 4808 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387386 4808 flags.go:64] FLAG: --rotate-certificates="false" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387391 4808 flags.go:64] FLAG: --rotate-server-certificates="false" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387396 4808 flags.go:64] FLAG: --runonce="false" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387401 4808 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387406 4808 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387412 4808 flags.go:64] FLAG: --seccomp-default="false" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387418 4808 flags.go:64] FLAG: --serialize-image-pulls="true" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387424 4808 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387430 4808 flags.go:64] FLAG: --storage-driver-db="cadvisor" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387435 4808 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387441 4808 flags.go:64] FLAG: --storage-driver-password="root" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387446 4808 flags.go:64] FLAG: --storage-driver-secure="false" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387452 4808 flags.go:64] FLAG: --storage-driver-table="stats" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387457 4808 flags.go:64] FLAG: --storage-driver-user="root" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387462 4808 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387468 4808 flags.go:64] FLAG: --sync-frequency="1m0s" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387473 4808 flags.go:64] FLAG: --system-cgroups="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387479 4808 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387487 4808 flags.go:64] FLAG: --system-reserved-cgroup="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387492 4808 flags.go:64] FLAG: --tls-cert-file="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387498 4808 flags.go:64] FLAG: --tls-cipher-suites="[]" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387505 4808 flags.go:64] FLAG: --tls-min-version="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387510 4808 flags.go:64] FLAG: --tls-private-key-file="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387516 4808 flags.go:64] FLAG: --topology-manager-policy="none" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387521 4808 flags.go:64] FLAG: --topology-manager-policy-options="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387526 4808 flags.go:64] FLAG: --topology-manager-scope="container" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387531 4808 flags.go:64] FLAG: --v="2" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387539 4808 flags.go:64] FLAG: --version="false" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387548 4808 flags.go:64] FLAG: --vmodule="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387555 4808 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.387560 4808 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387711 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387719 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387725 4808 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387731 4808 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387736 4808 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387740 4808 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387745 4808 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387750 4808 feature_gate.go:330] unrecognized feature gate: Example Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387754 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387759 4808 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387763 4808 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387768 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387772 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387778 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387783 4808 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387789 4808 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387795 4808 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387801 4808 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387806 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387811 4808 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387815 4808 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387820 4808 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387825 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387829 4808 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387834 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387838 4808 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387843 4808 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387848 4808 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387852 4808 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387857 4808 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387861 4808 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387866 4808 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387871 4808 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387875 4808 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387879 4808 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387885 4808 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387890 4808 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387895 4808 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387900 4808 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387904 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387909 4808 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387913 4808 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387918 4808 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387923 4808 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387927 4808 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387932 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387937 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387941 4808 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387946 4808 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387952 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387957 4808 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387962 4808 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387968 4808 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387974 4808 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387980 4808 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387985 4808 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387990 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387994 4808 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.387999 4808 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.388003 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.388008 4808 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.388012 4808 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.388018 4808 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.388024 4808 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.388031 4808 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.388036 4808 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.388041 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.388053 4808 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.388059 4808 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.388064 4808 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.388070 4808 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.388085 4808 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.398234 4808 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.398304 4808 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398437 4808 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398455 4808 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398461 4808 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398468 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398476 4808 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398483 4808 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398488 4808 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398494 4808 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398499 4808 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398505 4808 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398513 4808 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398519 4808 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398525 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398531 4808 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398537 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398544 4808 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398549 4808 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398554 4808 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398560 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398565 4808 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398570 4808 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398574 4808 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398579 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398585 4808 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398590 4808 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398596 4808 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398602 4808 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398608 4808 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398614 4808 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398620 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398626 4808 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398631 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398638 4808 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398643 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398651 4808 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398660 4808 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398667 4808 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398672 4808 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398678 4808 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398685 4808 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398692 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398697 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398702 4808 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398708 4808 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398713 4808 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398720 4808 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398725 4808 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398730 4808 feature_gate.go:330] unrecognized feature gate: Example Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398736 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398742 4808 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398747 4808 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398754 4808 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398762 4808 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398768 4808 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398773 4808 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398781 4808 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398789 4808 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398795 4808 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398800 4808 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398806 4808 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398810 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398816 4808 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398821 4808 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398826 4808 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398831 4808 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398836 4808 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398842 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398847 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398852 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398857 4808 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.398863 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.398872 4808 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399085 4808 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399096 4808 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399103 4808 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399108 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399115 4808 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399120 4808 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399125 4808 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399133 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399138 4808 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399143 4808 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399148 4808 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399152 4808 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399156 4808 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399163 4808 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399170 4808 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399174 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399179 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399185 4808 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399191 4808 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399197 4808 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399202 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399206 4808 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399211 4808 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399216 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399220 4808 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399225 4808 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399230 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399234 4808 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399257 4808 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399262 4808 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399268 4808 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399273 4808 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399278 4808 feature_gate.go:330] unrecognized feature gate: Example Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399282 4808 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399287 4808 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399292 4808 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399297 4808 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399303 4808 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399308 4808 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399312 4808 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399317 4808 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399321 4808 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399326 4808 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399330 4808 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399335 4808 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399339 4808 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399344 4808 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399350 4808 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399356 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399361 4808 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399368 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399373 4808 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399378 4808 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399383 4808 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399388 4808 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399393 4808 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399398 4808 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399403 4808 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399407 4808 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399413 4808 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399418 4808 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399423 4808 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399429 4808 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399437 4808 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399443 4808 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399449 4808 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399454 4808 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399460 4808 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399465 4808 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399471 4808 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.399476 4808 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.399484 4808 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.399935 4808 server.go:940] "Client rotation is on, will bootstrap in background" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.402939 4808 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.403046 4808 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.404027 4808 server.go:997] "Starting client certificate rotation" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.404060 4808 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.404509 4808 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-20 09:46:13.044441153 +0000 UTC Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.404588 4808 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.410754 4808 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 21 17:56:05 crc kubenswrapper[4808]: E0121 17:56:05.413661 4808 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.129.56.102:6443: connect: connection refused" logger="UnhandledError" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.413673 4808 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.422803 4808 log.go:25] "Validated CRI v1 runtime API" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.438665 4808 log.go:25] "Validated CRI v1 image API" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.440810 4808 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.443022 4808 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-01-21-17-50-16-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.443138 4808 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.458155 4808 manager.go:217] Machine: {Timestamp:2026-01-21 17:56:05.456988312 +0000 UTC m=+0.187499217 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654132736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:91b163f8-1521-490f-afb6-c160deb5beca BootID:7acf3d6e-f41a-4724-b064-e0293f86b6ec Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730829824 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827068416 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:78:06:33 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:78:06:33 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:6d:7f:0d Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:b1:21:cc Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:76:73:9e Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:46:e8:96 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ae:c3:da:dc:44:26 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:c2:1d:07:cc:94:d4 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654132736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.458415 4808 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.458624 4808 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.458910 4808 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.459073 4808 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.459103 4808 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.459339 4808 topology_manager.go:138] "Creating topology manager with none policy" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.459354 4808 container_manager_linux.go:303] "Creating device plugin manager" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.459559 4808 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.459607 4808 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.459872 4808 state_mem.go:36] "Initialized new in-memory state store" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.459952 4808 server.go:1245] "Using root directory" path="/var/lib/kubelet" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.461003 4808 kubelet.go:418] "Attempting to sync node with API server" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.461134 4808 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.461173 4808 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.461193 4808 kubelet.go:324] "Adding apiserver pod source" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.461207 4808 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.463195 4808 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.463470 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.102:6443: connect: connection refused Jan 21 17:56:05 crc kubenswrapper[4808]: E0121 17:56:05.463565 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.102:6443: connect: connection refused" logger="UnhandledError" Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.463555 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.102:6443: connect: connection refused Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.463601 4808 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Jan 21 17:56:05 crc kubenswrapper[4808]: E0121 17:56:05.463632 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.102:6443: connect: connection refused" logger="UnhandledError" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.464506 4808 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.465094 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.465121 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.465131 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.465139 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.465154 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.465163 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.465172 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.465186 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.465196 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.465205 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.465217 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.465225 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.465429 4808 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.465952 4808 server.go:1280] "Started kubelet" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.466644 4808 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.466639 4808 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.467229 4808 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.102:6443: connect: connection refused Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.467409 4808 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 21 17:56:05 crc systemd[1]: Started Kubernetes Kubelet. Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.468519 4808 server.go:460] "Adding debug handlers to kubelet server" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.468764 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.469002 4808 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.469123 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 10:26:46.123542943 +0000 UTC Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.469690 4808 volume_manager.go:287] "The desired_state_of_world populator starts" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.469709 4808 volume_manager.go:289] "Starting Kubelet Volume Manager" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.469814 4808 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 21 17:56:05 crc kubenswrapper[4808]: E0121 17:56:05.469481 4808 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.102:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188cd0a619cf1c06 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-21 17:56:05.465914374 +0000 UTC m=+0.196425279,LastTimestamp:2026-01-21 17:56:05.465914374 +0000 UTC m=+0.196425279,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 21 17:56:05 crc kubenswrapper[4808]: E0121 17:56:05.470587 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.102:6443: connect: connection refused" interval="200ms" Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.471205 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.102:6443: connect: connection refused Jan 21 17:56:05 crc kubenswrapper[4808]: E0121 17:56:05.471303 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.102:6443: connect: connection refused" logger="UnhandledError" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.471531 4808 factory.go:55] Registering systemd factory Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.471569 4808 factory.go:221] Registration of the systemd container factory successfully Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.471923 4808 factory.go:153] Registering CRI-O factory Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.471937 4808 factory.go:221] Registration of the crio container factory successfully Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.472046 4808 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.472081 4808 factory.go:103] Registering Raw factory Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.472099 4808 manager.go:1196] Started watching for new ooms in manager Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.474908 4808 manager.go:319] Starting recovery of all containers Jan 21 17:56:05 crc kubenswrapper[4808]: E0121 17:56:05.477452 4808 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.493848 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.493898 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.493912 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.493924 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.493937 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.493970 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.493980 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.493991 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494002 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494059 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494072 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494082 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494092 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494103 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494113 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494123 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494135 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494146 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494177 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494186 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494195 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494205 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494215 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494225 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494235 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494259 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494272 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494312 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494323 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494333 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494343 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494353 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494365 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494377 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494401 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494411 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494421 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494432 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494442 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494470 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494482 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494493 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.494503 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495318 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495361 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495374 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495387 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495399 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495414 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495426 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495438 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495449 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495468 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495507 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495523 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495534 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495546 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495557 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495568 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495579 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495592 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495603 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495617 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495628 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495640 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495650 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495660 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495671 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495681 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495693 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495705 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495716 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495727 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495739 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495749 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495762 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495772 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495785 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495795 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495805 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495818 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495828 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495867 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495882 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495892 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495902 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495912 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495921 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495931 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495941 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495950 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495960 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495972 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495982 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.495991 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496001 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496013 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496024 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496034 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496074 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496085 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496096 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496111 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496123 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496140 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496152 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496164 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496176 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496188 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496201 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496214 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496233 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496262 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496275 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496285 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496297 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496307 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496317 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496361 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496371 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496381 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496391 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496401 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496411 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496421 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496433 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.496443 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497171 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497204 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497227 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497312 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497327 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497340 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497355 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497366 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497378 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497393 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497411 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497423 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497436 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497451 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497465 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497477 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497489 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497505 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497516 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497530 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497543 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497556 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497571 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497607 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497620 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497632 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497644 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497656 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497668 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497690 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497701 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497712 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497725 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497738 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497750 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497763 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497775 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497786 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497796 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497812 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497822 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497834 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497846 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497856 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497868 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497880 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497890 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497900 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497911 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497923 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497934 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497944 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497954 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497964 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497975 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497986 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.497997 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498008 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498018 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498029 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498040 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498057 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498104 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498117 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498130 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498142 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498153 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498165 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498176 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498187 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498198 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498209 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498222 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498232 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498261 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498272 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498283 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498294 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498307 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498896 4808 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498918 4808 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498930 4808 reconstruct.go:97] "Volume reconstruction finished" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.498938 4808 reconciler.go:26] "Reconciler: start to sync state" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.499137 4808 manager.go:324] Recovery completed Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.509016 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.511196 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.511267 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.511280 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.512006 4808 cpu_manager.go:225] "Starting CPU manager" policy="none" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.512021 4808 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.512111 4808 state_mem.go:36] "Initialized new in-memory state store" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.516017 4808 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.518072 4808 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.519331 4808 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.519403 4808 kubelet.go:2335] "Starting kubelet main sync loop" Jan 21 17:56:05 crc kubenswrapper[4808]: E0121 17:56:05.519487 4808 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.520206 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.102:6443: connect: connection refused Jan 21 17:56:05 crc kubenswrapper[4808]: E0121 17:56:05.520349 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.102:6443: connect: connection refused" logger="UnhandledError" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.523698 4808 policy_none.go:49] "None policy: Start" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.525539 4808 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.525613 4808 state_mem.go:35] "Initializing new in-memory state store" Jan 21 17:56:05 crc kubenswrapper[4808]: E0121 17:56:05.579324 4808 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.582941 4808 manager.go:334] "Starting Device Plugin manager" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.583014 4808 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.583035 4808 server.go:79] "Starting device plugin registration server" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.583574 4808 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.583606 4808 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.583955 4808 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.584100 4808 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.584111 4808 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 21 17:56:05 crc kubenswrapper[4808]: E0121 17:56:05.591719 4808 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.619846 4808 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.619988 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.621201 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.621294 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.621311 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.621449 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.621748 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.621796 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.622274 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.622346 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.622362 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.622613 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.622654 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.622699 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.622711 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.622731 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.622733 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.623492 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.623532 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.623546 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.623808 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.623842 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.623853 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.623959 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.624233 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.624333 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.624657 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.624681 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.624702 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.624818 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.625027 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.625101 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.626345 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.626378 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.626392 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.626532 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.626557 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.626558 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.626591 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.626610 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.626569 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.626877 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.626915 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.628001 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.628036 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.628050 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:05 crc kubenswrapper[4808]: E0121 17:56:05.671577 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.102:6443: connect: connection refused" interval="400ms" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.684029 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.685507 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.685537 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.685546 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.685568 4808 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 21 17:56:05 crc kubenswrapper[4808]: E0121 17:56:05.685946 4808 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.102:6443: connect: connection refused" node="crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.701555 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.701597 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.701633 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.701658 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.701679 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.701757 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.701791 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.701810 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.701833 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.701848 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.701903 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.701936 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.701959 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.702029 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.702083 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803163 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803231 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803290 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803319 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803353 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803385 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803414 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803446 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803461 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803527 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803477 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803553 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803542 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803607 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803589 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803469 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803612 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803624 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803584 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803658 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803754 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803790 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803823 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803846 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803854 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803867 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803904 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803947 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.803942 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.804009 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.887093 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.888920 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.888987 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.889006 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.889048 4808 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 21 17:56:05 crc kubenswrapper[4808]: E0121 17:56:05.889707 4808 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.102:6443: connect: connection refused" node="crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.953188 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: W0121 17:56:05.990362 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-f3fe9a822be2a6240f4f817094cb519740434ce2ac0026480a6f2fe848d77fdf WatchSource:0}: Error finding container f3fe9a822be2a6240f4f817094cb519740434ce2ac0026480a6f2fe848d77fdf: Status 404 returned error can't find the container with id f3fe9a822be2a6240f4f817094cb519740434ce2ac0026480a6f2fe848d77fdf Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.990644 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 21 17:56:05 crc kubenswrapper[4808]: I0121 17:56:05.999083 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.006137 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 17:56:06 crc kubenswrapper[4808]: W0121 17:56:06.016781 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-0cd044e6ccdde00ebcee4dd8dc8a5d6600988e350bffbffac6c3e668ed387982 WatchSource:0}: Error finding container 0cd044e6ccdde00ebcee4dd8dc8a5d6600988e350bffbffac6c3e668ed387982: Status 404 returned error can't find the container with id 0cd044e6ccdde00ebcee4dd8dc8a5d6600988e350bffbffac6c3e668ed387982 Jan 21 17:56:06 crc kubenswrapper[4808]: W0121 17:56:06.024544 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-06a5d36b510d15da43707bfdb142e4590f19d80d0d9e1e46142c205b50f93ff0 WatchSource:0}: Error finding container 06a5d36b510d15da43707bfdb142e4590f19d80d0d9e1e46142c205b50f93ff0: Status 404 returned error can't find the container with id 06a5d36b510d15da43707bfdb142e4590f19d80d0d9e1e46142c205b50f93ff0 Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.029913 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 17:56:06 crc kubenswrapper[4808]: W0121 17:56:06.030292 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-57fde7c363d9aa51e6876ab46048f6d7579385791d68e3187954dff9ae76fc44 WatchSource:0}: Error finding container 57fde7c363d9aa51e6876ab46048f6d7579385791d68e3187954dff9ae76fc44: Status 404 returned error can't find the container with id 57fde7c363d9aa51e6876ab46048f6d7579385791d68e3187954dff9ae76fc44 Jan 21 17:56:06 crc kubenswrapper[4808]: E0121 17:56:06.073157 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.102:6443: connect: connection refused" interval="800ms" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.290108 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.291730 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.291774 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.291783 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.291807 4808 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 21 17:56:06 crc kubenswrapper[4808]: E0121 17:56:06.292357 4808 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.102:6443: connect: connection refused" node="crc" Jan 21 17:56:06 crc kubenswrapper[4808]: W0121 17:56:06.388712 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.102:6443: connect: connection refused Jan 21 17:56:06 crc kubenswrapper[4808]: E0121 17:56:06.388817 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.102:6443: connect: connection refused" logger="UnhandledError" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.468026 4808 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.102:6443: connect: connection refused Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.470046 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 15:44:07.113536152 +0000 UTC Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.524562 4808 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="4522d0820456caa3b2de83a5b240e5f904ab45e4326101a0e35fefd7b80ffbba" exitCode=0 Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.524651 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"4522d0820456caa3b2de83a5b240e5f904ab45e4326101a0e35fefd7b80ffbba"} Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.524764 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"f3fe9a822be2a6240f4f817094cb519740434ce2ac0026480a6f2fe848d77fdf"} Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.524862 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.526211 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.526279 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.526314 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.527424 4808 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6" exitCode=0 Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.527491 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6"} Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.527512 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"47db615b9e9940582cf02e9aaadacfadcf7a2a749b5fee84fb68fe9ea4acf80a"} Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.527582 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.528189 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.528215 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.528227 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.529446 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92"} Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.529475 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"57fde7c363d9aa51e6876ab46048f6d7579385791d68e3187954dff9ae76fc44"} Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.531330 4808 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1" exitCode=0 Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.531392 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1"} Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.531413 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"06a5d36b510d15da43707bfdb142e4590f19d80d0d9e1e46142c205b50f93ff0"} Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.531488 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.532453 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.532479 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.532490 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.534021 4808 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2a0beb110ecaec1b20d9534ba045b026f6f81d5843432b3c339a4d483e8db907" exitCode=0 Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.534062 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2a0beb110ecaec1b20d9534ba045b026f6f81d5843432b3c339a4d483e8db907"} Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.534083 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0cd044e6ccdde00ebcee4dd8dc8a5d6600988e350bffbffac6c3e668ed387982"} Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.534177 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.534831 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.535211 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.535255 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.535269 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.535763 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.535800 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:06 crc kubenswrapper[4808]: I0121 17:56:06.535828 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:06 crc kubenswrapper[4808]: W0121 17:56:06.739616 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.102:6443: connect: connection refused Jan 21 17:56:06 crc kubenswrapper[4808]: E0121 17:56:06.739681 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.102:6443: connect: connection refused" logger="UnhandledError" Jan 21 17:56:06 crc kubenswrapper[4808]: W0121 17:56:06.806091 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.102:6443: connect: connection refused Jan 21 17:56:06 crc kubenswrapper[4808]: E0121 17:56:06.806176 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.102:6443: connect: connection refused" logger="UnhandledError" Jan 21 17:56:06 crc kubenswrapper[4808]: W0121 17:56:06.833148 4808 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.102:6443: connect: connection refused Jan 21 17:56:06 crc kubenswrapper[4808]: E0121 17:56:06.833254 4808 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.102:6443: connect: connection refused" logger="UnhandledError" Jan 21 17:56:06 crc kubenswrapper[4808]: E0121 17:56:06.874017 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.102:6443: connect: connection refused" interval="1.6s" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.093479 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.095192 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.095260 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.095274 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.095303 4808 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 21 17:56:07 crc kubenswrapper[4808]: E0121 17:56:07.095839 4808 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.102:6443: connect: connection refused" node="crc" Jan 21 17:56:07 crc kubenswrapper[4808]: E0121 17:56:07.140059 4808 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.102:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188cd0a619cf1c06 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-21 17:56:05.465914374 +0000 UTC m=+0.196425279,LastTimestamp:2026-01-21 17:56:05.465914374 +0000 UTC m=+0.196425279,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.417798 4808 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.470524 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 05:20:21.693194003 +0000 UTC Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.541621 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"94585d6e08b85c96da3d90042ca20e0a194e7488185ef798b4c2dcc2c2948897"} Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.541668 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"59727192447c6eef4b045ac5a9a0c0a7bba9c0f2672d46ff1fc9b2b58c022279"} Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.541678 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"65ea8da4528954ea712538a48e6e1c6a1fe893873742ad120f6300ee32f6c95b"} Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.541761 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.542737 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.542766 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.542775 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.546581 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e"} Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.546642 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd"} Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.546652 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.546658 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a"} Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.547602 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.547643 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.547656 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.549965 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7"} Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.549992 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6"} Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.550004 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1"} Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.550018 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159"} Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.551690 4808 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="df9b6e87aaad39e9f7ad32d0ecb1460ccd9bea92984d5c781ded9f5fd1fb3fe0" exitCode=0 Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.551742 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"df9b6e87aaad39e9f7ad32d0ecb1460ccd9bea92984d5c781ded9f5fd1fb3fe0"} Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.551847 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.552609 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.552632 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.552643 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.554293 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"4b73ce1747c3ddbb5b62b30dcbe4a7f8e5c8196d388af47a7b882f6a411cadbe"} Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.554359 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.555387 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.555410 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:07 crc kubenswrapper[4808]: I0121 17:56:07.555420 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.037019 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.471172 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 12:36:11.678402112 +0000 UTC Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.560889 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7"} Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.560961 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.561781 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.561826 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.561844 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.563506 4808 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="207f21fb18b2f4ce24e583ecdc278b1ffdbc2088180d1b01675dd260857a4ece" exitCode=0 Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.563579 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"207f21fb18b2f4ce24e583ecdc278b1ffdbc2088180d1b01675dd260857a4ece"} Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.563594 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.563729 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.564461 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.564497 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.564515 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.564515 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.564548 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.564558 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.696203 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.697828 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.697880 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.697890 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:08 crc kubenswrapper[4808]: I0121 17:56:08.697922 4808 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 21 17:56:09 crc kubenswrapper[4808]: I0121 17:56:09.116039 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 17:56:09 crc kubenswrapper[4808]: I0121 17:56:09.122611 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 17:56:09 crc kubenswrapper[4808]: I0121 17:56:09.471593 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 17:48:56.924184984 +0000 UTC Jan 21 17:56:09 crc kubenswrapper[4808]: I0121 17:56:09.571371 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"93668cebf0cfa604e697766701302dde52d52174b7e6520459f55b94b36ec8f3"} Jan 21 17:56:09 crc kubenswrapper[4808]: I0121 17:56:09.571437 4808 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 21 17:56:09 crc kubenswrapper[4808]: I0121 17:56:09.571490 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:09 crc kubenswrapper[4808]: I0121 17:56:09.571445 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c976dc9fcd39e0eb7f489e08a80c70af0b8a28cd35996c8b3f9e9788f47b3034"} Jan 21 17:56:09 crc kubenswrapper[4808]: I0121 17:56:09.571598 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:09 crc kubenswrapper[4808]: I0121 17:56:09.571590 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9af59d705eedee87d5f3230fabdca74d1ce715ad5d8c8e6d2c13d6fee8dcb4e7"} Jan 21 17:56:09 crc kubenswrapper[4808]: I0121 17:56:09.571736 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fa998e75b5ed636c6ccb207d654e83b8b92c76cba9d36143cc770c9c752b81a9"} Jan 21 17:56:09 crc kubenswrapper[4808]: I0121 17:56:09.572415 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:09 crc kubenswrapper[4808]: I0121 17:56:09.572463 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:09 crc kubenswrapper[4808]: I0121 17:56:09.572482 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:09 crc kubenswrapper[4808]: I0121 17:56:09.572963 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:09 crc kubenswrapper[4808]: I0121 17:56:09.573009 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:09 crc kubenswrapper[4808]: I0121 17:56:09.573022 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:09 crc kubenswrapper[4808]: I0121 17:56:09.642343 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 17:56:09 crc kubenswrapper[4808]: I0121 17:56:09.928945 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 17:56:10 crc kubenswrapper[4808]: I0121 17:56:10.011965 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 17:56:10 crc kubenswrapper[4808]: I0121 17:56:10.214012 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 17:56:10 crc kubenswrapper[4808]: I0121 17:56:10.472055 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 15:18:10.115114382 +0000 UTC Jan 21 17:56:10 crc kubenswrapper[4808]: I0121 17:56:10.577786 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"cb700754f9313898ada246adcd5c398b43d4a2acdb83919980e3e22a0839a2a5"} Jan 21 17:56:10 crc kubenswrapper[4808]: I0121 17:56:10.577981 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:10 crc kubenswrapper[4808]: I0121 17:56:10.578079 4808 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 21 17:56:10 crc kubenswrapper[4808]: I0121 17:56:10.578186 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:10 crc kubenswrapper[4808]: I0121 17:56:10.577995 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:10 crc kubenswrapper[4808]: I0121 17:56:10.579480 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:10 crc kubenswrapper[4808]: I0121 17:56:10.579564 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:10 crc kubenswrapper[4808]: I0121 17:56:10.579611 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:10 crc kubenswrapper[4808]: I0121 17:56:10.579546 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:10 crc kubenswrapper[4808]: I0121 17:56:10.579647 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:10 crc kubenswrapper[4808]: I0121 17:56:10.579657 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:10 crc kubenswrapper[4808]: I0121 17:56:10.579629 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:10 crc kubenswrapper[4808]: I0121 17:56:10.579585 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:10 crc kubenswrapper[4808]: I0121 17:56:10.579850 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.037841 4808 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.037951 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.190436 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.190865 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.193003 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.193094 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.193116 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.472587 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 21:16:08.573190175 +0000 UTC Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.580036 4808 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.580097 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.580053 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.580054 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.580962 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.581003 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.581100 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.581997 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.582102 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.582188 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.582022 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.582348 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:11 crc kubenswrapper[4808]: I0121 17:56:11.582358 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:12 crc kubenswrapper[4808]: I0121 17:56:12.473222 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 04:46:18.46861086 +0000 UTC Jan 21 17:56:13 crc kubenswrapper[4808]: I0121 17:56:13.173081 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Jan 21 17:56:13 crc kubenswrapper[4808]: I0121 17:56:13.173483 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:13 crc kubenswrapper[4808]: I0121 17:56:13.175221 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:13 crc kubenswrapper[4808]: I0121 17:56:13.175303 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:13 crc kubenswrapper[4808]: I0121 17:56:13.175321 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:13 crc kubenswrapper[4808]: I0121 17:56:13.474830 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 23:20:47.324168797 +0000 UTC Jan 21 17:56:14 crc kubenswrapper[4808]: I0121 17:56:14.475348 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 09:23:01.11117832 +0000 UTC Jan 21 17:56:15 crc kubenswrapper[4808]: I0121 17:56:15.397623 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 17:56:15 crc kubenswrapper[4808]: I0121 17:56:15.397827 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:15 crc kubenswrapper[4808]: I0121 17:56:15.398977 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:15 crc kubenswrapper[4808]: I0121 17:56:15.399011 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:15 crc kubenswrapper[4808]: I0121 17:56:15.399022 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:15 crc kubenswrapper[4808]: I0121 17:56:15.476021 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 12:21:13.731010221 +0000 UTC Jan 21 17:56:15 crc kubenswrapper[4808]: E0121 17:56:15.591825 4808 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 21 17:56:16 crc kubenswrapper[4808]: I0121 17:56:16.476651 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 15:53:01.366594212 +0000 UTC Jan 21 17:56:17 crc kubenswrapper[4808]: E0121 17:56:17.420531 4808 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Jan 21 17:56:17 crc kubenswrapper[4808]: I0121 17:56:17.468625 4808 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Jan 21 17:56:17 crc kubenswrapper[4808]: I0121 17:56:17.477080 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 02:34:34.087239285 +0000 UTC Jan 21 17:56:18 crc kubenswrapper[4808]: I0121 17:56:18.467787 4808 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 21 17:56:18 crc kubenswrapper[4808]: I0121 17:56:18.468146 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 21 17:56:18 crc kubenswrapper[4808]: I0121 17:56:18.471926 4808 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 21 17:56:18 crc kubenswrapper[4808]: I0121 17:56:18.472002 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 21 17:56:18 crc kubenswrapper[4808]: I0121 17:56:18.477232 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 19:45:14.992367608 +0000 UTC Jan 21 17:56:19 crc kubenswrapper[4808]: I0121 17:56:19.477767 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 05:36:41.862040151 +0000 UTC Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.020963 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.021188 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.022916 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.022985 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.023011 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.223906 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.224161 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.225799 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.225849 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.225859 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.230455 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.470471 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.470774 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.472806 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.472890 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.472910 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.478762 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 21:18:59.57628067 +0000 UTC Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.506334 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.600902 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.601008 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.603071 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.603149 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.603173 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.603707 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.603869 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.603899 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:20 crc kubenswrapper[4808]: I0121 17:56:20.623595 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Jan 21 17:56:21 crc kubenswrapper[4808]: I0121 17:56:21.037997 4808 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 21 17:56:21 crc kubenswrapper[4808]: I0121 17:56:21.038115 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 21 17:56:21 crc kubenswrapper[4808]: I0121 17:56:21.479716 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 17:21:54.934198371 +0000 UTC Jan 21 17:56:21 crc kubenswrapper[4808]: I0121 17:56:21.603260 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:21 crc kubenswrapper[4808]: I0121 17:56:21.604429 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:21 crc kubenswrapper[4808]: I0121 17:56:21.604491 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:21 crc kubenswrapper[4808]: I0121 17:56:21.604515 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:21 crc kubenswrapper[4808]: I0121 17:56:21.754766 4808 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 21 17:56:21 crc kubenswrapper[4808]: I0121 17:56:21.771519 4808 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 21 17:56:22 crc kubenswrapper[4808]: I0121 17:56:22.480620 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 16:29:35.473033257 +0000 UTC Jan 21 17:56:23 crc kubenswrapper[4808]: E0121 17:56:23.436772 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.439120 4808 trace.go:236] Trace[515885274]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Jan-2026 17:56:09.768) (total time: 13670ms): Jan 21 17:56:23 crc kubenswrapper[4808]: Trace[515885274]: ---"Objects listed" error: 13669ms (17:56:23.438) Jan 21 17:56:23 crc kubenswrapper[4808]: Trace[515885274]: [13.670150217s] [13.670150217s] END Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.439186 4808 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.439324 4808 trace.go:236] Trace[472240334]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Jan-2026 17:56:09.876) (total time: 13562ms): Jan 21 17:56:23 crc kubenswrapper[4808]: Trace[472240334]: ---"Objects listed" error: 13562ms (17:56:23.438) Jan 21 17:56:23 crc kubenswrapper[4808]: Trace[472240334]: [13.562810999s] [13.562810999s] END Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.439360 4808 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.440626 4808 trace.go:236] Trace[904584526]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Jan-2026 17:56:09.418) (total time: 14022ms): Jan 21 17:56:23 crc kubenswrapper[4808]: Trace[904584526]: ---"Objects listed" error: 14022ms (17:56:23.440) Jan 21 17:56:23 crc kubenswrapper[4808]: Trace[904584526]: [14.022304867s] [14.022304867s] END Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.440652 4808 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 21 17:56:23 crc kubenswrapper[4808]: E0121 17:56:23.441119 4808 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.442409 4808 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.444480 4808 trace.go:236] Trace[1113032208]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Jan-2026 17:56:09.332) (total time: 14111ms): Jan 21 17:56:23 crc kubenswrapper[4808]: Trace[1113032208]: ---"Objects listed" error: 14111ms (17:56:23.444) Jan 21 17:56:23 crc kubenswrapper[4808]: Trace[1113032208]: [14.111750074s] [14.111750074s] END Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.444521 4808 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.472655 4808 apiserver.go:52] "Watching apiserver" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.476227 4808 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.476723 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.477172 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.477350 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:23 crc kubenswrapper[4808]: E0121 17:56:23.477483 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.477506 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.477537 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 17:56:23 crc kubenswrapper[4808]: E0121 17:56:23.477566 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.478136 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.478196 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 17:56:23 crc kubenswrapper[4808]: E0121 17:56:23.478208 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.480567 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.480590 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.480571 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.480673 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.480789 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 06:35:00.388676229 +0000 UTC Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.481025 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.481385 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.481691 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.481808 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.482058 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.499500 4808 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:50744->192.168.126.11:17697: read: connection reset by peer" start-of-body= Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.499557 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:50744->192.168.126.11:17697: read: connection reset by peer" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.499394 4808 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:50740->192.168.126.11:17697: read: connection reset by peer" start-of-body= Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.499633 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:50740->192.168.126.11:17697: read: connection reset by peer" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.499960 4808 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.500018 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.507491 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.518579 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.534405 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.547540 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.557792 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.568302 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.570755 4808 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.578174 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.588992 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.600837 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.608846 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.610854 4808 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7" exitCode=255 Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.610890 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7"} Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.628045 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.643645 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.643700 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.643753 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.643776 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.643804 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.643827 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.643847 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.643867 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.643888 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.643930 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.643954 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.643976 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.643991 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.643997 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644031 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644056 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644127 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644132 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644154 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644191 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644193 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644223 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644266 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644291 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644306 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644320 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644361 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644385 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644428 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644441 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644467 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644482 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644490 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644530 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644557 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644584 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644626 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644702 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644738 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644136 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644769 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644856 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644910 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644950 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644992 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645022 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645054 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645079 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645098 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644947 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644969 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644966 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644985 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644997 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645017 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: E0121 17:56:23.645132 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:56:24.145096223 +0000 UTC m=+18.875607328 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645479 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645485 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645504 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645528 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645554 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645549 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645578 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645597 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645602 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645650 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645651 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645681 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645691 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645696 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645709 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645743 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645769 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645787 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645792 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645806 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645830 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645850 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645867 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645883 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645898 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645947 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645966 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645990 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646013 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646038 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646058 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646074 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646091 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646112 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646179 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646198 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646214 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646230 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646260 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646276 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646292 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646308 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646322 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646340 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646357 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646373 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646387 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646402 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646416 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646431 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646448 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646464 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646480 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646496 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646513 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646528 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646573 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646589 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646605 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646623 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646639 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646656 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646671 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646687 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646703 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646718 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646734 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646753 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646769 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646786 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646800 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646822 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646838 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645825 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645938 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645156 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645164 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645362 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645381 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645395 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645984 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646005 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646000 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646015 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646019 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646138 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.645152 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646184 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646340 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646460 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646475 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646533 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.647058 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.647067 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.647123 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.647188 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.647563 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.647597 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.647631 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.647715 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.644265 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649069 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649380 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649465 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649392 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646680 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646819 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649524 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649548 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649566 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649628 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649648 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649664 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649681 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649698 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649715 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649731 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649748 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649764 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649782 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649802 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649820 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649848 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649874 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649898 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649918 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649943 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649949 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649967 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.649989 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650010 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650031 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650049 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650064 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650080 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650096 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650112 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650129 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650144 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650160 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650176 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650192 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650208 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650223 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650254 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650274 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650290 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650308 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650332 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650352 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650372 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650392 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650408 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650426 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650442 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650457 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650473 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650489 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650505 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650522 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650557 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650574 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650592 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650608 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650624 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650640 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650657 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650676 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650696 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650743 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650760 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650779 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650796 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650812 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650828 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650845 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650862 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650878 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650893 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650908 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650923 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650938 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650954 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650970 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.650987 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651005 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651027 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651052 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651076 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651099 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651121 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651146 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651165 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651142 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651182 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651341 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651386 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651416 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651453 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651484 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651564 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651630 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651660 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651664 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651690 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651746 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651775 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651824 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651852 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651872 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651893 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651912 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651932 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651964 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.651983 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652103 4808 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652120 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652133 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652144 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652155 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652157 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652170 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652185 4808 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652197 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652209 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652220 4808 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652251 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652266 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652277 4808 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652287 4808 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652297 4808 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652307 4808 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652320 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652333 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652347 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652359 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652372 4808 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652384 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652394 4808 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652404 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652414 4808 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652424 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652434 4808 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652444 4808 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652457 4808 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652471 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652482 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652492 4808 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652503 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652521 4808 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652539 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652553 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652566 4808 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652582 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652600 4808 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652625 4808 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652635 4808 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652645 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652656 4808 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652668 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652680 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652691 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652702 4808 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652712 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652722 4808 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652736 4808 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652750 4808 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652765 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652779 4808 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652794 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652806 4808 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652816 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652827 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652837 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652847 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652858 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652594 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.652725 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.653029 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.654063 4808 scope.go:117] "RemoveContainer" containerID="15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.653228 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.646578 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.653547 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.656800 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.657034 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.657272 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.658078 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.658165 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.659979 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.660271 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.660522 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.660539 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.661146 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.661561 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.662136 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.662228 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.662480 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.664345 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.664454 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.665346 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.665667 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.667303 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.667541 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.667619 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.667652 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.667808 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.667904 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.668774 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.668653 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.679620 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.680149 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.680237 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.680338 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.680686 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.680679 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.681105 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.681194 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.681403 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.681639 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.681756 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.681717 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.682150 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.681977 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.682505 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.682536 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.682818 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.682828 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.682931 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.682950 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.683234 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.683321 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.683609 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.683651 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.683897 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.684308 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.684689 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.684963 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.685203 4808 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.690036 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.690602 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.680457 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.691064 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.691735 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.691906 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.692756 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.692894 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.692960 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: E0121 17:56:23.693167 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 17:56:23 crc kubenswrapper[4808]: E0121 17:56:23.693341 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:24.193301288 +0000 UTC m=+18.923812183 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.693883 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.694803 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.698050 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.699518 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.700134 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.700459 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.701750 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.702226 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.702402 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: E0121 17:56:23.702672 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 17:56:23 crc kubenswrapper[4808]: E0121 17:56:23.702793 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:24.202771096 +0000 UTC m=+18.933281981 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.703107 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.703606 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.703913 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.704030 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.704433 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.709769 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.710415 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.710638 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.710866 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.711222 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.715399 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.715493 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.716367 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.716814 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.716966 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.717046 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.717278 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.718561 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.719014 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.719534 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.720100 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.722085 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.722414 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: E0121 17:56:23.724523 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.724530 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: E0121 17:56:23.724557 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 17:56:23 crc kubenswrapper[4808]: E0121 17:56:23.724577 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:23 crc kubenswrapper[4808]: E0121 17:56:23.724694 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:24.224660875 +0000 UTC m=+18.955171760 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.725191 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.726672 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.728392 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.729890 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.732837 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 17:56:23 crc kubenswrapper[4808]: E0121 17:56:23.734405 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 17:56:23 crc kubenswrapper[4808]: E0121 17:56:23.734496 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 17:56:23 crc kubenswrapper[4808]: E0121 17:56:23.734596 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.734615 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 17:56:23 crc kubenswrapper[4808]: E0121 17:56:23.734691 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:24.234671844 +0000 UTC m=+18.965182729 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.739696 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.740096 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.740677 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.741338 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.741563 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.741518 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.742042 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.742113 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.746773 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.746836 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.746901 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.747389 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.748228 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.748679 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.748866 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.748955 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.748854 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.749136 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.749348 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.749495 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.749598 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.749833 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.750391 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.751497 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.751600 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.752078 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.752310 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.752480 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.752995 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753428 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753502 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753571 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753586 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753601 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753615 4808 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753628 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753639 4808 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753651 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753662 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753673 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753684 4808 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753694 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753703 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753714 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753723 4808 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753732 4808 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753742 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753755 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753765 4808 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753775 4808 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753786 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753792 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753797 4808 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753833 4808 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753845 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753858 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753869 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753880 4808 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753890 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753901 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753911 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753921 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753931 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753942 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753952 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753962 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753972 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753982 4808 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753994 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754004 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754016 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754027 4808 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754037 4808 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754046 4808 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754058 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754068 4808 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754079 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754089 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754099 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754110 4808 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754121 4808 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754132 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754143 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754153 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754165 4808 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754176 4808 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754187 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754198 4808 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754207 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754217 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754227 4808 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754255 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.753657 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754268 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754334 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754349 4808 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754360 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754368 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754378 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754387 4808 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754397 4808 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754406 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754414 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754423 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754431 4808 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754438 4808 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754447 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754477 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754486 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754496 4808 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754505 4808 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754515 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754525 4808 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754534 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754543 4808 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754552 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754560 4808 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754569 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754578 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754587 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754595 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754593 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754603 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754642 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754654 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754667 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754679 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754692 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754703 4808 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754719 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754732 4808 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754743 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.754755 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756056 4808 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756070 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756089 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756099 4808 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756108 4808 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756117 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756126 4808 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756134 4808 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756143 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756154 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756164 4808 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756172 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756181 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756190 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756201 4808 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756210 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756218 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756226 4808 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756234 4808 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756258 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756267 4808 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756275 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756284 4808 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756294 4808 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756303 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756312 4808 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756446 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756674 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756852 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.756867 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.758107 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.763154 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.767400 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.780665 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.792489 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.798867 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.806625 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 17:56:23 crc kubenswrapper[4808]: W0121 17:56:23.813885 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-b0768f411069063e9e85071802294d56c63c54c53bbd1794898442e4e06e95ea WatchSource:0}: Error finding container b0768f411069063e9e85071802294d56c63c54c53bbd1794898442e4e06e95ea: Status 404 returned error can't find the container with id b0768f411069063e9e85071802294d56c63c54c53bbd1794898442e4e06e95ea Jan 21 17:56:23 crc kubenswrapper[4808]: W0121 17:56:23.814685 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-3d9aa802bde0dc7861814978208993f4d4fa88c655351a9cd6ca232ad8752af0 WatchSource:0}: Error finding container 3d9aa802bde0dc7861814978208993f4d4fa88c655351a9cd6ca232ad8752af0: Status 404 returned error can't find the container with id 3d9aa802bde0dc7861814978208993f4d4fa88c655351a9cd6ca232ad8752af0 Jan 21 17:56:23 crc kubenswrapper[4808]: W0121 17:56:23.835649 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-6f5287f181f7bc5a0bb50bc83c4131a13b7413ca536af4813881fc58fb424aaf WatchSource:0}: Error finding container 6f5287f181f7bc5a0bb50bc83c4131a13b7413ca536af4813881fc58fb424aaf: Status 404 returned error can't find the container with id 6f5287f181f7bc5a0bb50bc83c4131a13b7413ca536af4813881fc58fb424aaf Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.856674 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.856712 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.856726 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.856740 4808 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.856751 4808 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.856764 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.856777 4808 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.856790 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:23 crc kubenswrapper[4808]: I0121 17:56:23.856801 4808 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.159496 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:56:24 crc kubenswrapper[4808]: E0121 17:56:24.159692 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:56:25.159664959 +0000 UTC m=+19.890175844 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.260373 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.260436 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.260478 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.260516 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:24 crc kubenswrapper[4808]: E0121 17:56:24.260561 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 17:56:24 crc kubenswrapper[4808]: E0121 17:56:24.260589 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 17:56:24 crc kubenswrapper[4808]: E0121 17:56:24.260603 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:24 crc kubenswrapper[4808]: E0121 17:56:24.260610 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 17:56:24 crc kubenswrapper[4808]: E0121 17:56:24.260637 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 17:56:24 crc kubenswrapper[4808]: E0121 17:56:24.260658 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:25.26063983 +0000 UTC m=+19.991150725 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:24 crc kubenswrapper[4808]: E0121 17:56:24.260660 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 17:56:24 crc kubenswrapper[4808]: E0121 17:56:24.260675 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:25.260667941 +0000 UTC m=+19.991178826 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 17:56:24 crc kubenswrapper[4808]: E0121 17:56:24.260680 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:24 crc kubenswrapper[4808]: E0121 17:56:24.260687 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 17:56:24 crc kubenswrapper[4808]: E0121 17:56:24.260730 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:25.260713032 +0000 UTC m=+19.991223957 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:24 crc kubenswrapper[4808]: E0121 17:56:24.260780 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:25.260761333 +0000 UTC m=+19.991272218 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.481141 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 03:53:21.215483972 +0000 UTC Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.614233 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"6f5287f181f7bc5a0bb50bc83c4131a13b7413ca536af4813881fc58fb424aaf"} Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.615959 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2"} Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.615991 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766"} Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.616000 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b0768f411069063e9e85071802294d56c63c54c53bbd1794898442e4e06e95ea"} Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.617431 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6"} Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.617452 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"3d9aa802bde0dc7861814978208993f4d4fa88c655351a9cd6ca232ad8752af0"} Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.619885 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.621873 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51"} Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.622420 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.640343 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:24Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.656514 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:24Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.677954 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:24Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.693772 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:24Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.710581 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:24Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.724539 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:24Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.735867 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:24Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.756938 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:24Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.780845 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:24Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.797990 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:24Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.810127 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:24Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.821765 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:24Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.836460 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:24Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:24 crc kubenswrapper[4808]: I0121 17:56:24.851627 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:24Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.170112 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:56:25 crc kubenswrapper[4808]: E0121 17:56:25.170425 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:56:27.170400679 +0000 UTC m=+21.900911564 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.271597 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.271640 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.271659 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.271676 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:25 crc kubenswrapper[4808]: E0121 17:56:25.271728 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 17:56:25 crc kubenswrapper[4808]: E0121 17:56:25.271753 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 17:56:25 crc kubenswrapper[4808]: E0121 17:56:25.271787 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:27.271770909 +0000 UTC m=+22.002281794 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 17:56:25 crc kubenswrapper[4808]: E0121 17:56:25.271799 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:27.271793759 +0000 UTC m=+22.002304644 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 17:56:25 crc kubenswrapper[4808]: E0121 17:56:25.271817 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 17:56:25 crc kubenswrapper[4808]: E0121 17:56:25.271827 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 17:56:25 crc kubenswrapper[4808]: E0121 17:56:25.271836 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:25 crc kubenswrapper[4808]: E0121 17:56:25.271858 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:27.271850651 +0000 UTC m=+22.002361536 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:25 crc kubenswrapper[4808]: E0121 17:56:25.271922 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 17:56:25 crc kubenswrapper[4808]: E0121 17:56:25.271982 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 17:56:25 crc kubenswrapper[4808]: E0121 17:56:25.272000 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:25 crc kubenswrapper[4808]: E0121 17:56:25.272074 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:27.272053795 +0000 UTC m=+22.002564730 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.482451 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 09:54:39.673658752 +0000 UTC Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.519700 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.519726 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.519753 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:25 crc kubenswrapper[4808]: E0121 17:56:25.519882 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:56:25 crc kubenswrapper[4808]: E0121 17:56:25.520013 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:56:25 crc kubenswrapper[4808]: E0121 17:56:25.520104 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.524949 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.525956 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.528173 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.529444 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.530225 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.530968 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.531745 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.532478 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.533354 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.534006 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.534616 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.535462 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.536054 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.536716 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.537310 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.537999 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.539394 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.539989 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.540929 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.543395 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.544303 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.544770 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.545372 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.545815 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.546528 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.546937 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.547563 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.548233 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.548801 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.549370 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.549840 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.550294 4808 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.550391 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.551762 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.552289 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.552673 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.556661 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.557576 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.557660 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.558353 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.559337 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.559987 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.560799 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.561388 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.562374 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.563317 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.563758 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.564312 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.565125 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.566472 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.566931 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.567381 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.568160 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.568654 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.569584 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.570039 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.575476 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.593622 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.615729 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.630038 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:25 crc kubenswrapper[4808]: I0121 17:56:25.645105 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.483549 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 08:41:45.170377724 +0000 UTC Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.641797 4808 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.644343 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.644377 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.644388 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.644454 4808 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.652564 4808 kubelet_node_status.go:115] "Node was previously registered" node="crc" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.652887 4808 kubelet_node_status.go:79] "Successfully registered node" node="crc" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.654163 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.654220 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.654295 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.654332 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.654385 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:26Z","lastTransitionTime":"2026-01-21T17:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:26 crc kubenswrapper[4808]: E0121 17:56:26.673377 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:26Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.676696 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.676723 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.676732 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.676770 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.676780 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:26Z","lastTransitionTime":"2026-01-21T17:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:26 crc kubenswrapper[4808]: E0121 17:56:26.688645 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:26Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.691730 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.691796 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.691808 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.691823 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.691848 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:26Z","lastTransitionTime":"2026-01-21T17:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:26 crc kubenswrapper[4808]: E0121 17:56:26.704355 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:26Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.709905 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.709944 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.709956 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.709973 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.709985 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:26Z","lastTransitionTime":"2026-01-21T17:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:26 crc kubenswrapper[4808]: E0121 17:56:26.720984 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:26Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.723885 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.723916 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.723962 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.723977 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.723987 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:26Z","lastTransitionTime":"2026-01-21T17:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:26 crc kubenswrapper[4808]: E0121 17:56:26.736758 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:26Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:26 crc kubenswrapper[4808]: E0121 17:56:26.737099 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.738470 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.738498 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.738511 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.738525 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.738537 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:26Z","lastTransitionTime":"2026-01-21T17:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.841413 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.841479 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.841502 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.841531 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.841555 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:26Z","lastTransitionTime":"2026-01-21T17:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.944462 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.944525 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.944535 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.944550 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:26 crc kubenswrapper[4808]: I0121 17:56:26.944560 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:26Z","lastTransitionTime":"2026-01-21T17:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.048755 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.048839 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.048867 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.048900 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.048924 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:27Z","lastTransitionTime":"2026-01-21T17:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.151046 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.151104 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.151119 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.151140 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.151153 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:27Z","lastTransitionTime":"2026-01-21T17:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.189436 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:56:27 crc kubenswrapper[4808]: E0121 17:56:27.189611 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:56:31.18959356 +0000 UTC m=+25.920104455 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.253359 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.253414 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.253429 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.253453 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.253470 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:27Z","lastTransitionTime":"2026-01-21T17:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.290691 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.290739 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.290771 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.290808 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:27 crc kubenswrapper[4808]: E0121 17:56:27.290864 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 17:56:27 crc kubenswrapper[4808]: E0121 17:56:27.290897 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 17:56:27 crc kubenswrapper[4808]: E0121 17:56:27.290904 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 17:56:27 crc kubenswrapper[4808]: E0121 17:56:27.290911 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:27 crc kubenswrapper[4808]: E0121 17:56:27.290966 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:31.290948769 +0000 UTC m=+26.021459674 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 17:56:27 crc kubenswrapper[4808]: E0121 17:56:27.290985 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:31.29097627 +0000 UTC m=+26.021487175 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:27 crc kubenswrapper[4808]: E0121 17:56:27.291002 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 17:56:27 crc kubenswrapper[4808]: E0121 17:56:27.291043 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:31.291023481 +0000 UTC m=+26.021534436 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 17:56:27 crc kubenswrapper[4808]: E0121 17:56:27.291067 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 17:56:27 crc kubenswrapper[4808]: E0121 17:56:27.291086 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 17:56:27 crc kubenswrapper[4808]: E0121 17:56:27.291102 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:27 crc kubenswrapper[4808]: E0121 17:56:27.291145 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:31.291133644 +0000 UTC m=+26.021644619 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.356775 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.356842 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.356881 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.356900 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.356912 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:27Z","lastTransitionTime":"2026-01-21T17:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.459342 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.459385 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.459395 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.459410 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.459420 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:27Z","lastTransitionTime":"2026-01-21T17:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.484873 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 23:16:31.90068229 +0000 UTC Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.520696 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.520696 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:27 crc kubenswrapper[4808]: E0121 17:56:27.520923 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.521000 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:27 crc kubenswrapper[4808]: E0121 17:56:27.521135 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:56:27 crc kubenswrapper[4808]: E0121 17:56:27.521317 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.562295 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.562327 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.562338 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.562353 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.562363 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:27Z","lastTransitionTime":"2026-01-21T17:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.630471 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb"} Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.653384 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.664382 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.664430 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.664443 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.664468 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.664480 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:27Z","lastTransitionTime":"2026-01-21T17:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.670625 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.687684 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.702231 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.717304 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.736867 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.746734 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.766582 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.766612 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.766619 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.766649 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.766661 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:27Z","lastTransitionTime":"2026-01-21T17:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.868696 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.868743 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.868758 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.868779 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.868789 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:27Z","lastTransitionTime":"2026-01-21T17:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.971794 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.971858 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.971873 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.971902 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:27 crc kubenswrapper[4808]: I0121 17:56:27.971917 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:27Z","lastTransitionTime":"2026-01-21T17:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.060046 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.065302 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.069894 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.074375 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.074419 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.074432 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.074451 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.074467 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:28Z","lastTransitionTime":"2026-01-21T17:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.075511 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:28Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.087812 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:28Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.098478 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:28Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.116067 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:28Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.126544 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:28Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.150261 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:28Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.166755 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:28Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.177196 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.177268 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.177285 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.177309 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.177327 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:28Z","lastTransitionTime":"2026-01-21T17:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.183285 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:28Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.202029 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:28Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.217175 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:28Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.231804 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:28Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.245193 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:28Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.261302 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:28Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.274317 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:28Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.279148 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.279186 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.279195 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.279209 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.279218 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:28Z","lastTransitionTime":"2026-01-21T17:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.289530 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:28Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.382982 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.383055 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.383076 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.383110 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.383137 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:28Z","lastTransitionTime":"2026-01-21T17:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.485442 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 23:58:47.325562526 +0000 UTC Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.485983 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.486045 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.486058 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.486074 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.486085 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:28Z","lastTransitionTime":"2026-01-21T17:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.588962 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.589004 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.589012 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.589029 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.589038 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:28Z","lastTransitionTime":"2026-01-21T17:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.691687 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.691756 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.691772 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.691797 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.691808 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:28Z","lastTransitionTime":"2026-01-21T17:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.794551 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.794586 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.794593 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.794607 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.794617 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:28Z","lastTransitionTime":"2026-01-21T17:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.811852 4808 csr.go:261] certificate signing request csr-p5pbs is approved, waiting to be issued Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.826181 4808 csr.go:257] certificate signing request csr-p5pbs is issued Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.896722 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.896759 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.896769 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.896787 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.896800 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:28Z","lastTransitionTime":"2026-01-21T17:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.999484 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.999561 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.999572 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.999592 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:28 crc kubenswrapper[4808]: I0121 17:56:28.999601 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:28Z","lastTransitionTime":"2026-01-21T17:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.101577 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.101619 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.101631 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.101648 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.101660 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:29Z","lastTransitionTime":"2026-01-21T17:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.203872 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.203920 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.203929 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.203944 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.203955 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:29Z","lastTransitionTime":"2026-01-21T17:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.306373 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.306410 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.306419 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.306437 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.306447 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:29Z","lastTransitionTime":"2026-01-21T17:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.408770 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.408815 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.408828 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.408846 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.408858 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:29Z","lastTransitionTime":"2026-01-21T17:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.485938 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 11:42:36.984939407 +0000 UTC Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.511620 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.511659 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.511670 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.511685 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.511698 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:29Z","lastTransitionTime":"2026-01-21T17:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.519637 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.519657 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:29 crc kubenswrapper[4808]: E0121 17:56:29.519726 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.519821 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:29 crc kubenswrapper[4808]: E0121 17:56:29.519924 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:56:29 crc kubenswrapper[4808]: E0121 17:56:29.519977 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.613664 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.613700 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.613711 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.613728 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.613741 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:29Z","lastTransitionTime":"2026-01-21T17:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.699482 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-sm4d2"] Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.699805 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-sm4d2" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.702363 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.702628 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.702907 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.706137 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-lgtv9"] Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.706540 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.707929 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.708763 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fc2fk"] Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.709080 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.709151 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.709274 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.709638 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.712907 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.713082 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.713094 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.713174 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.713513 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.713605 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.713619 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.713619 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.715432 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.715470 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.715482 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.715500 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.715512 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:29Z","lastTransitionTime":"2026-01-21T17:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.717292 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-4kc58"] Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.717990 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-z8fp9"] Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.718174 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.718328 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: W0121 17:56:29.719451 4808 reflector.go:561] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": failed to list *v1.Secret: secrets "multus-ancillary-tools-dockercfg-vnmsz" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Jan 21 17:56:29 crc kubenswrapper[4808]: W0121 17:56:29.719466 4808 reflector.go:561] object-"openshift-multus"/"default-dockercfg-2q5b6": failed to list *v1.Secret: secrets "default-dockercfg-2q5b6" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Jan 21 17:56:29 crc kubenswrapper[4808]: W0121 17:56:29.719481 4808 reflector.go:561] object-"openshift-multus"/"multus-daemon-config": failed to list *v1.ConfigMap: configmaps "multus-daemon-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Jan 21 17:56:29 crc kubenswrapper[4808]: E0121 17:56:29.719486 4808 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-vnmsz\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"multus-ancillary-tools-dockercfg-vnmsz\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 21 17:56:29 crc kubenswrapper[4808]: E0121 17:56:29.719489 4808 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"default-dockercfg-2q5b6\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"default-dockercfg-2q5b6\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 21 17:56:29 crc kubenswrapper[4808]: W0121 17:56:29.719499 4808 reflector.go:561] object-"openshift-multus"/"default-cni-sysctl-allowlist": failed to list *v1.ConfigMap: configmaps "default-cni-sysctl-allowlist" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Jan 21 17:56:29 crc kubenswrapper[4808]: E0121 17:56:29.719517 4808 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"default-cni-sysctl-allowlist\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 21 17:56:29 crc kubenswrapper[4808]: E0121 17:56:29.719515 4808 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-daemon-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"multus-daemon-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 21 17:56:29 crc kubenswrapper[4808]: W0121 17:56:29.719875 4808 reflector.go:561] object-"openshift-multus"/"cni-copy-resources": failed to list *v1.ConfigMap: configmaps "cni-copy-resources" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Jan 21 17:56:29 crc kubenswrapper[4808]: E0121 17:56:29.719914 4808 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"cni-copy-resources\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"cni-copy-resources\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 21 17:56:29 crc kubenswrapper[4808]: W0121 17:56:29.720327 4808 reflector.go:561] object-"openshift-multus"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Jan 21 17:56:29 crc kubenswrapper[4808]: E0121 17:56:29.720351 4808 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 21 17:56:29 crc kubenswrapper[4808]: W0121 17:56:29.720378 4808 reflector.go:561] object-"openshift-multus"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Jan 21 17:56:29 crc kubenswrapper[4808]: E0121 17:56:29.720401 4808 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.723621 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:29Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.736105 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:29Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.749466 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:29Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.764224 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:29Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.785107 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:29Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.808396 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:29Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.815942 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-var-lib-openvswitch\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.815988 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-run-openvswitch\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.816014 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-run-ovn-kubernetes\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.816040 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d76d1c92-28d2-4476-8af9-d95cc779425e-rootfs\") pod \"machine-config-daemon-lgtv9\" (UID: \"d76d1c92-28d2-4476-8af9-d95cc779425e\") " pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.816097 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bb13a20f-f4ab-412c-8165-5b93edc79628-env-overrides\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.816133 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/538d1d44-6c67-450a-809c-d9008128ec9f-cni-binary-copy\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.816154 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4ffed385-f2d1-4043-9323-561451d85fad-os-release\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.816178 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-slash\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.816198 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-host-run-multus-certs\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.816217 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bb13a20f-f4ab-412c-8165-5b93edc79628-ovn-node-metrics-cert\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.816252 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp5bs\" (UniqueName: \"kubernetes.io/projected/d76d1c92-28d2-4476-8af9-d95cc779425e-kube-api-access-xp5bs\") pod \"machine-config-daemon-lgtv9\" (UID: \"d76d1c92-28d2-4476-8af9-d95cc779425e\") " pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.816275 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-host-run-k8s-cni-cncf-io\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.816583 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/4ffed385-f2d1-4043-9323-561451d85fad-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.816624 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-systemd-units\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.816647 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-run-netns\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.816672 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnpck\" (UniqueName: \"kubernetes.io/projected/538d1d44-6c67-450a-809c-d9008128ec9f-kube-api-access-fnpck\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.816696 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4ffed385-f2d1-4043-9323-561451d85fad-cni-binary-copy\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.816720 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d76d1c92-28d2-4476-8af9-d95cc779425e-mcd-auth-proxy-config\") pod \"machine-config-daemon-lgtv9\" (UID: \"d76d1c92-28d2-4476-8af9-d95cc779425e\") " pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.816741 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-etc-kubernetes\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.816787 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/538d1d44-6c67-450a-809c-d9008128ec9f-multus-daemon-config\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.816810 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-log-socket\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.817016 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx7ms\" (UniqueName: \"kubernetes.io/projected/b6c02096-5729-4118-9b7a-0b3722acf2b6-kube-api-access-kx7ms\") pod \"node-resolver-sm4d2\" (UID: \"b6c02096-5729-4118-9b7a-0b3722acf2b6\") " pod="openshift-dns/node-resolver-sm4d2" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.817051 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-cni-netd\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.817080 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2824\" (UniqueName: \"kubernetes.io/projected/4ffed385-f2d1-4043-9323-561451d85fad-kube-api-access-k2824\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.817120 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d76d1c92-28d2-4476-8af9-d95cc779425e-proxy-tls\") pod \"machine-config-daemon-lgtv9\" (UID: \"d76d1c92-28d2-4476-8af9-d95cc779425e\") " pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.817145 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-multus-cni-dir\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.817170 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4ffed385-f2d1-4043-9323-561451d85fad-cnibin\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.817193 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4ffed385-f2d1-4043-9323-561451d85fad-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.817215 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-cni-bin\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.817351 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-run-ovn\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.817381 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bb13a20f-f4ab-412c-8165-5b93edc79628-ovnkube-config\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.817406 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-multus-conf-dir\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.817495 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.817523 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwp7h\" (UniqueName: \"kubernetes.io/projected/bb13a20f-f4ab-412c-8165-5b93edc79628-kube-api-access-zwp7h\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.817576 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-cnibin\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.817604 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ffed385-f2d1-4043-9323-561451d85fad-system-cni-dir\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.817726 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-kubelet\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.817748 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-run-systemd\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.817789 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bb13a20f-f4ab-412c-8165-5b93edc79628-ovnkube-script-lib\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.818032 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-os-release\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.818357 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-host-var-lib-cni-multus\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.818390 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-host-run-netns\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.818475 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-etc-openvswitch\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.818508 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-system-cni-dir\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.818547 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-node-log\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.818582 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-host-var-lib-kubelet\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.818608 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-multus-socket-dir-parent\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.818644 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-host-var-lib-cni-bin\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.818657 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.818665 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-hostroot\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.818684 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.818697 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.818715 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.818719 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b6c02096-5729-4118-9b7a-0b3722acf2b6-hosts-file\") pod \"node-resolver-sm4d2\" (UID: \"b6c02096-5729-4118-9b7a-0b3722acf2b6\") " pod="openshift-dns/node-resolver-sm4d2" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.818731 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:29Z","lastTransitionTime":"2026-01-21T17:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.826536 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:29Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.827341 4808 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-01-21 17:51:28 +0000 UTC, rotation deadline is 2026-11-29 23:09:21.585342905 +0000 UTC Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.827396 4808 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7493h12m51.757948569s for next certificate rotation Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.838445 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:29Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.854918 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:29Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.875156 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:29Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.887780 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:29Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.905400 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:29Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.919955 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bb13a20f-f4ab-412c-8165-5b93edc79628-ovnkube-script-lib\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.919993 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-os-release\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920010 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-host-var-lib-cni-multus\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920027 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-host-run-netns\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920042 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-etc-openvswitch\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920056 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-system-cni-dir\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920071 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-node-log\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920085 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-host-var-lib-kubelet\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920088 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-host-var-lib-cni-multus\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920100 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-multus-socket-dir-parent\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920114 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-host-var-lib-cni-bin\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920127 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-hostroot\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920142 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b6c02096-5729-4118-9b7a-0b3722acf2b6-hosts-file\") pod \"node-resolver-sm4d2\" (UID: \"b6c02096-5729-4118-9b7a-0b3722acf2b6\") " pod="openshift-dns/node-resolver-sm4d2" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920155 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-var-lib-openvswitch\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920168 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-run-openvswitch\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920170 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-os-release\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920182 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-run-ovn-kubernetes\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920196 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d76d1c92-28d2-4476-8af9-d95cc779425e-rootfs\") pod \"machine-config-daemon-lgtv9\" (UID: \"d76d1c92-28d2-4476-8af9-d95cc779425e\") " pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920221 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bb13a20f-f4ab-412c-8165-5b93edc79628-env-overrides\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920236 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/538d1d44-6c67-450a-809c-d9008128ec9f-cni-binary-copy\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920268 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4ffed385-f2d1-4043-9323-561451d85fad-os-release\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920282 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-slash\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920296 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-host-run-multus-certs\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920299 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-host-run-netns\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920311 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bb13a20f-f4ab-412c-8165-5b93edc79628-ovn-node-metrics-cert\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920325 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp5bs\" (UniqueName: \"kubernetes.io/projected/d76d1c92-28d2-4476-8af9-d95cc779425e-kube-api-access-xp5bs\") pod \"machine-config-daemon-lgtv9\" (UID: \"d76d1c92-28d2-4476-8af9-d95cc779425e\") " pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920340 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-host-run-k8s-cni-cncf-io\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920351 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-etc-openvswitch\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920355 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/4ffed385-f2d1-4043-9323-561451d85fad-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920377 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-systemd-units\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920397 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-run-netns\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920410 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-hostroot\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920437 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnpck\" (UniqueName: \"kubernetes.io/projected/538d1d44-6c67-450a-809c-d9008128ec9f-kube-api-access-fnpck\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920446 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b6c02096-5729-4118-9b7a-0b3722acf2b6-hosts-file\") pod \"node-resolver-sm4d2\" (UID: \"b6c02096-5729-4118-9b7a-0b3722acf2b6\") " pod="openshift-dns/node-resolver-sm4d2" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920461 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4ffed385-f2d1-4043-9323-561451d85fad-cni-binary-copy\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920467 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-var-lib-openvswitch\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920495 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-run-openvswitch\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920484 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d76d1c92-28d2-4476-8af9-d95cc779425e-mcd-auth-proxy-config\") pod \"machine-config-daemon-lgtv9\" (UID: \"d76d1c92-28d2-4476-8af9-d95cc779425e\") " pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920516 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-run-ovn-kubernetes\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920520 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-etc-kubernetes\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920537 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d76d1c92-28d2-4476-8af9-d95cc779425e-rootfs\") pod \"machine-config-daemon-lgtv9\" (UID: \"d76d1c92-28d2-4476-8af9-d95cc779425e\") " pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920541 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/538d1d44-6c67-450a-809c-d9008128ec9f-multus-daemon-config\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920555 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-log-socket\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920585 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx7ms\" (UniqueName: \"kubernetes.io/projected/b6c02096-5729-4118-9b7a-0b3722acf2b6-kube-api-access-kx7ms\") pod \"node-resolver-sm4d2\" (UID: \"b6c02096-5729-4118-9b7a-0b3722acf2b6\") " pod="openshift-dns/node-resolver-sm4d2" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920598 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-cni-netd\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920614 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2824\" (UniqueName: \"kubernetes.io/projected/4ffed385-f2d1-4043-9323-561451d85fad-kube-api-access-k2824\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920637 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d76d1c92-28d2-4476-8af9-d95cc779425e-proxy-tls\") pod \"machine-config-daemon-lgtv9\" (UID: \"d76d1c92-28d2-4476-8af9-d95cc779425e\") " pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920666 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-multus-cni-dir\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920681 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4ffed385-f2d1-4043-9323-561451d85fad-cnibin\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920697 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4ffed385-f2d1-4043-9323-561451d85fad-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920710 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-cni-bin\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920740 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-run-ovn\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920753 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bb13a20f-f4ab-412c-8165-5b93edc79628-ovnkube-config\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920766 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-multus-conf-dir\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920781 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920279 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-host-var-lib-cni-bin\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920796 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwp7h\" (UniqueName: \"kubernetes.io/projected/bb13a20f-f4ab-412c-8165-5b93edc79628-kube-api-access-zwp7h\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920840 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-cnibin\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920855 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ffed385-f2d1-4043-9323-561451d85fad-system-cni-dir\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920857 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4ffed385-f2d1-4043-9323-561451d85fad-os-release\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920868 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-kubelet\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920883 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-slash\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920882 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-run-systemd\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920905 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bb13a20f-f4ab-412c-8165-5b93edc79628-ovnkube-script-lib\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920928 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bb13a20f-f4ab-412c-8165-5b93edc79628-env-overrides\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.920947 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-run-systemd\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921047 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-host-run-k8s-cni-cncf-io\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921099 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-system-cni-dir\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921129 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-node-log\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921145 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-host-run-multus-certs\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921167 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-host-var-lib-kubelet\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921194 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-multus-socket-dir-parent\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921210 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-systemd-units\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921213 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921223 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-run-netns\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921236 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921263 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921281 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921296 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:29Z","lastTransitionTime":"2026-01-21T17:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921565 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-etc-kubernetes\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921628 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4ffed385-f2d1-4043-9323-561451d85fad-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921636 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-log-socket\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921781 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-cni-netd\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921826 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4ffed385-f2d1-4043-9323-561451d85fad-cnibin\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921847 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d76d1c92-28d2-4476-8af9-d95cc779425e-mcd-auth-proxy-config\") pod \"machine-config-daemon-lgtv9\" (UID: \"d76d1c92-28d2-4476-8af9-d95cc779425e\") " pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921885 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-cni-bin\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921918 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-multus-cni-dir\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.921927 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-run-ovn\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.922002 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-cnibin\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.922036 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ffed385-f2d1-4043-9323-561451d85fad-system-cni-dir\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.922084 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-kubelet\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.922117 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/538d1d44-6c67-450a-809c-d9008128ec9f-multus-conf-dir\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.922118 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.922547 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bb13a20f-f4ab-412c-8165-5b93edc79628-ovnkube-config\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.924847 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:29Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.928774 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bb13a20f-f4ab-412c-8165-5b93edc79628-ovn-node-metrics-cert\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.935901 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d76d1c92-28d2-4476-8af9-d95cc779425e-proxy-tls\") pod \"machine-config-daemon-lgtv9\" (UID: \"d76d1c92-28d2-4476-8af9-d95cc779425e\") " pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.936128 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp5bs\" (UniqueName: \"kubernetes.io/projected/d76d1c92-28d2-4476-8af9-d95cc779425e-kube-api-access-xp5bs\") pod \"machine-config-daemon-lgtv9\" (UID: \"d76d1c92-28d2-4476-8af9-d95cc779425e\") " pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.936988 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwp7h\" (UniqueName: \"kubernetes.io/projected/bb13a20f-f4ab-412c-8165-5b93edc79628-kube-api-access-zwp7h\") pod \"ovnkube-node-fc2fk\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.937113 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx7ms\" (UniqueName: \"kubernetes.io/projected/b6c02096-5729-4118-9b7a-0b3722acf2b6-kube-api-access-kx7ms\") pod \"node-resolver-sm4d2\" (UID: \"b6c02096-5729-4118-9b7a-0b3722acf2b6\") " pod="openshift-dns/node-resolver-sm4d2" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.937802 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:29Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.950119 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:29Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.961301 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:29Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.972033 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:29Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:29 crc kubenswrapper[4808]: I0121 17:56:29.989455 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:29Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.000591 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:29Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.011832 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-sm4d2" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.012060 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.022078 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.022455 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: W0121 17:56:30.022272 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6c02096_5729_4118_9b7a_0b3722acf2b6.slice/crio-208652f49fe3357408d8e3f6263780eeb69dcfe015c4801ca66f680d50261098 WatchSource:0}: Error finding container 208652f49fe3357408d8e3f6263780eeb69dcfe015c4801ca66f680d50261098: Status 404 returned error can't find the container with id 208652f49fe3357408d8e3f6263780eeb69dcfe015c4801ca66f680d50261098 Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.023000 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.023027 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.023038 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.023053 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.023064 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:30Z","lastTransitionTime":"2026-01-21T17:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.029718 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.032599 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: W0121 17:56:30.036097 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd76d1c92_28d2_4476_8af9_d95cc779425e.slice/crio-296c9e0d522468facd8d86dfab5e4db511e767f96a44d2456f852409b19a23e3 WatchSource:0}: Error finding container 296c9e0d522468facd8d86dfab5e4db511e767f96a44d2456f852409b19a23e3: Status 404 returned error can't find the container with id 296c9e0d522468facd8d86dfab5e4db511e767f96a44d2456f852409b19a23e3 Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.125127 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.125163 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.125174 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.125195 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.125204 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:30Z","lastTransitionTime":"2026-01-21T17:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.227184 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.227219 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.227230 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.227258 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.227270 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:30Z","lastTransitionTime":"2026-01-21T17:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.329327 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.329375 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.329390 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.329411 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.329430 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:30Z","lastTransitionTime":"2026-01-21T17:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.432751 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.433105 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.433115 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.433131 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.433140 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:30Z","lastTransitionTime":"2026-01-21T17:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.489379 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 18:03:50.159772703 +0000 UTC Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.535872 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.535910 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.535919 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.535935 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.535945 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:30Z","lastTransitionTime":"2026-01-21T17:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.637962 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.638034 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.638046 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.638062 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.638078 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:30Z","lastTransitionTime":"2026-01-21T17:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.643411 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-sm4d2" event={"ID":"b6c02096-5729-4118-9b7a-0b3722acf2b6","Type":"ContainerStarted","Data":"a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec"} Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.643464 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-sm4d2" event={"ID":"b6c02096-5729-4118-9b7a-0b3722acf2b6","Type":"ContainerStarted","Data":"208652f49fe3357408d8e3f6263780eeb69dcfe015c4801ca66f680d50261098"} Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.645842 4808 generic.go:334] "Generic (PLEG): container finished" podID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerID="d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653" exitCode=0 Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.645913 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerDied","Data":"d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653"} Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.645960 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerStarted","Data":"bdceb85fb9701f1d5c24bc358c55d9e3cbc44b1259f1a3f401679dc516c72136"} Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.649286 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerStarted","Data":"9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1"} Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.649328 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerStarted","Data":"d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb"} Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.649338 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerStarted","Data":"296c9e0d522468facd8d86dfab5e4db511e767f96a44d2456f852409b19a23e3"} Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.664675 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.679186 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.693326 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.707484 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.720306 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.732255 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.740201 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.740229 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.740237 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.740272 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.740286 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:30Z","lastTransitionTime":"2026-01-21T17:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.743886 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.761236 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.772790 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.787478 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.799399 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.810336 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.818199 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.820745 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.821951 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/538d1d44-6c67-450a-809c-d9008128ec9f-cni-binary-copy\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.822281 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4ffed385-f2d1-4043-9323-561451d85fad-cni-binary-copy\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.831709 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.843736 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.843772 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.843783 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.843799 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.843808 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:30Z","lastTransitionTime":"2026-01-21T17:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.844883 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.856368 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.868305 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.881806 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.893099 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.903411 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.908165 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.911140 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/4ffed385-f2d1-4043-9323-561451d85fad-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.914483 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 21 17:56:30 crc kubenswrapper[4808]: E0121 17:56:30.922176 4808 configmap.go:193] Couldn't get configMap openshift-multus/multus-daemon-config: failed to sync configmap cache: timed out waiting for the condition Jan 21 17:56:30 crc kubenswrapper[4808]: E0121 17:56:30.922289 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/538d1d44-6c67-450a-809c-d9008128ec9f-multus-daemon-config podName:538d1d44-6c67-450a-809c-d9008128ec9f nodeName:}" failed. No retries permitted until 2026-01-21 17:56:31.422264909 +0000 UTC m=+26.152775804 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "multus-daemon-config" (UniqueName: "kubernetes.io/configmap/538d1d44-6c67-450a-809c-d9008128ec9f-multus-daemon-config") pod "multus-z8fp9" (UID: "538d1d44-6c67-450a-809c-d9008128ec9f") : failed to sync configmap cache: timed out waiting for the condition Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.922278 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: E0121 17:56:30.936959 4808 projected.go:288] Couldn't get configMap openshift-multus/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Jan 21 17:56:30 crc kubenswrapper[4808]: E0121 17:56:30.936965 4808 projected.go:288] Couldn't get configMap openshift-multus/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.939149 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.945807 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.946020 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.946158 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.946286 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.946410 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:30Z","lastTransitionTime":"2026-01-21T17:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.950363 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.951726 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.963950 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.967168 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.976663 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:30 crc kubenswrapper[4808]: E0121 17:56:30.977721 4808 projected.go:194] Error preparing data for projected volume kube-api-access-k2824 for pod openshift-multus/multus-additional-cni-plugins-4kc58: failed to sync configmap cache: timed out waiting for the condition Jan 21 17:56:30 crc kubenswrapper[4808]: E0121 17:56:30.977825 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4ffed385-f2d1-4043-9323-561451d85fad-kube-api-access-k2824 podName:4ffed385-f2d1-4043-9323-561451d85fad nodeName:}" failed. No retries permitted until 2026-01-21 17:56:31.477802805 +0000 UTC m=+26.208313700 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-k2824" (UniqueName: "kubernetes.io/projected/4ffed385-f2d1-4043-9323-561451d85fad-kube-api-access-k2824") pod "multus-additional-cni-plugins-4kc58" (UID: "4ffed385-f2d1-4043-9323-561451d85fad") : failed to sync configmap cache: timed out waiting for the condition Jan 21 17:56:30 crc kubenswrapper[4808]: E0121 17:56:30.977735 4808 projected.go:194] Error preparing data for projected volume kube-api-access-fnpck for pod openshift-multus/multus-z8fp9: failed to sync configmap cache: timed out waiting for the condition Jan 21 17:56:30 crc kubenswrapper[4808]: E0121 17:56:30.977913 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/538d1d44-6c67-450a-809c-d9008128ec9f-kube-api-access-fnpck podName:538d1d44-6c67-450a-809c-d9008128ec9f nodeName:}" failed. No retries permitted until 2026-01-21 17:56:31.477890807 +0000 UTC m=+26.208401752 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-fnpck" (UniqueName: "kubernetes.io/projected/538d1d44-6c67-450a-809c-d9008128ec9f-kube-api-access-fnpck") pod "multus-z8fp9" (UID: "538d1d44-6c67-450a-809c-d9008128ec9f") : failed to sync configmap cache: timed out waiting for the condition Jan 21 17:56:30 crc kubenswrapper[4808]: I0121 17:56:30.988042 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:30Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.048844 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.048887 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.048899 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.048914 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.048924 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:31Z","lastTransitionTime":"2026-01-21T17:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.151312 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.151373 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.151385 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.151399 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.151409 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:31Z","lastTransitionTime":"2026-01-21T17:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.181841 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.234133 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:56:31 crc kubenswrapper[4808]: E0121 17:56:31.234372 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:56:39.234357103 +0000 UTC m=+33.964867988 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.255534 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.255561 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.255569 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.255582 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.255590 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:31Z","lastTransitionTime":"2026-01-21T17:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.258043 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.335830 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:31 crc kubenswrapper[4808]: E0121 17:56:31.336035 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 17:56:31 crc kubenswrapper[4808]: E0121 17:56:31.336291 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 17:56:31 crc kubenswrapper[4808]: E0121 17:56:31.336307 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.336265 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:31 crc kubenswrapper[4808]: E0121 17:56:31.336360 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:39.336345126 +0000 UTC m=+34.066856011 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.336474 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.336563 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:31 crc kubenswrapper[4808]: E0121 17:56:31.336588 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 17:56:31 crc kubenswrapper[4808]: E0121 17:56:31.336666 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 17:56:31 crc kubenswrapper[4808]: E0121 17:56:31.336671 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:39.336653693 +0000 UTC m=+34.067164578 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 17:56:31 crc kubenswrapper[4808]: E0121 17:56:31.336755 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:39.336741104 +0000 UTC m=+34.067251989 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 17:56:31 crc kubenswrapper[4808]: E0121 17:56:31.336861 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 17:56:31 crc kubenswrapper[4808]: E0121 17:56:31.336941 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 17:56:31 crc kubenswrapper[4808]: E0121 17:56:31.337004 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:31 crc kubenswrapper[4808]: E0121 17:56:31.337107 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:39.337092682 +0000 UTC m=+34.067603557 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.357779 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.357817 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.357826 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.357841 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.357852 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:31Z","lastTransitionTime":"2026-01-21T17:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.437341 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/538d1d44-6c67-450a-809c-d9008128ec9f-multus-daemon-config\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.438200 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/538d1d44-6c67-450a-809c-d9008128ec9f-multus-daemon-config\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.459418 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.459455 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.459463 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.459478 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.459489 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:31Z","lastTransitionTime":"2026-01-21T17:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.489712 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 01:05:50.850735318 +0000 UTC Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.519595 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.519595 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:31 crc kubenswrapper[4808]: E0121 17:56:31.519725 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:56:31 crc kubenswrapper[4808]: E0121 17:56:31.519773 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.519988 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:31 crc kubenswrapper[4808]: E0121 17:56:31.520161 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.538930 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnpck\" (UniqueName: \"kubernetes.io/projected/538d1d44-6c67-450a-809c-d9008128ec9f-kube-api-access-fnpck\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.538999 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2824\" (UniqueName: \"kubernetes.io/projected/4ffed385-f2d1-4043-9323-561451d85fad-kube-api-access-k2824\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.545835 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2824\" (UniqueName: \"kubernetes.io/projected/4ffed385-f2d1-4043-9323-561451d85fad-kube-api-access-k2824\") pod \"multus-additional-cni-plugins-4kc58\" (UID: \"4ffed385-f2d1-4043-9323-561451d85fad\") " pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.545892 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnpck\" (UniqueName: \"kubernetes.io/projected/538d1d44-6c67-450a-809c-d9008128ec9f-kube-api-access-fnpck\") pod \"multus-z8fp9\" (UID: \"538d1d44-6c67-450a-809c-d9008128ec9f\") " pod="openshift-multus/multus-z8fp9" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.561352 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.561387 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.561396 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.561410 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.561420 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:31Z","lastTransitionTime":"2026-01-21T17:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.655720 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerStarted","Data":"1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f"} Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.655786 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerStarted","Data":"25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd"} Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.655799 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerStarted","Data":"6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53"} Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.655809 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerStarted","Data":"b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739"} Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.655818 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerStarted","Data":"c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53"} Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.655829 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerStarted","Data":"5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053"} Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.663323 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.663375 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.663387 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.663404 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.663417 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:31Z","lastTransitionTime":"2026-01-21T17:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.766013 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.766054 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.766084 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.766100 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.766110 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:31Z","lastTransitionTime":"2026-01-21T17:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.837960 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-4kc58" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.844612 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-z8fp9" Jan 21 17:56:31 crc kubenswrapper[4808]: W0121 17:56:31.849981 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ffed385_f2d1_4043_9323_561451d85fad.slice/crio-abde942d1c668e59c3f519498a86161da8fcf3d5cb8b26c62a66f51aa7f9644a WatchSource:0}: Error finding container abde942d1c668e59c3f519498a86161da8fcf3d5cb8b26c62a66f51aa7f9644a: Status 404 returned error can't find the container with id abde942d1c668e59c3f519498a86161da8fcf3d5cb8b26c62a66f51aa7f9644a Jan 21 17:56:31 crc kubenswrapper[4808]: W0121 17:56:31.859033 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod538d1d44_6c67_450a_809c_d9008128ec9f.slice/crio-e8460b6c236faf92bca86b5f54e80d5bffd9b53c9e3f2812cebca96a192a3c93 WatchSource:0}: Error finding container e8460b6c236faf92bca86b5f54e80d5bffd9b53c9e3f2812cebca96a192a3c93: Status 404 returned error can't find the container with id e8460b6c236faf92bca86b5f54e80d5bffd9b53c9e3f2812cebca96a192a3c93 Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.871737 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.871947 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.872075 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.872175 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.872276 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:31Z","lastTransitionTime":"2026-01-21T17:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.976980 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.977015 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.977025 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.977044 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:31 crc kubenswrapper[4808]: I0121 17:56:31.977055 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:31Z","lastTransitionTime":"2026-01-21T17:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.079204 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.079235 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.079259 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.079274 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.079285 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:32Z","lastTransitionTime":"2026-01-21T17:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.182439 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.182465 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.182474 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.182487 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.182495 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:32Z","lastTransitionTime":"2026-01-21T17:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.284666 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.284695 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.284704 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.284717 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.284728 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:32Z","lastTransitionTime":"2026-01-21T17:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.387647 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.387681 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.387690 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.387705 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.387717 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:32Z","lastTransitionTime":"2026-01-21T17:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.510409 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 08:42:37.759525127 +0000 UTC Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.512197 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.512231 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.512253 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.512268 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.512276 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:32Z","lastTransitionTime":"2026-01-21T17:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.614712 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.614752 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.614763 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.614780 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.614792 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:32Z","lastTransitionTime":"2026-01-21T17:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.649857 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-28p27"] Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.650165 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-28p27" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.651651 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.652115 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.652222 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.652804 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.665419 4808 generic.go:334] "Generic (PLEG): container finished" podID="4ffed385-f2d1-4043-9323-561451d85fad" containerID="9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970" exitCode=0 Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.665512 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" event={"ID":"4ffed385-f2d1-4043-9323-561451d85fad","Type":"ContainerDied","Data":"9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970"} Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.665571 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" event={"ID":"4ffed385-f2d1-4043-9323-561451d85fad","Type":"ContainerStarted","Data":"abde942d1c668e59c3f519498a86161da8fcf3d5cb8b26c62a66f51aa7f9644a"} Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.667280 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.669294 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8fp9" event={"ID":"538d1d44-6c67-450a-809c-d9008128ec9f","Type":"ContainerStarted","Data":"33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397"} Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.669334 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8fp9" event={"ID":"538d1d44-6c67-450a-809c-d9008128ec9f","Type":"ContainerStarted","Data":"e8460b6c236faf92bca86b5f54e80d5bffd9b53c9e3f2812cebca96a192a3c93"} Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.685930 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.705666 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.717553 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.718469 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.718492 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.718504 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.718521 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.718534 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:32Z","lastTransitionTime":"2026-01-21T17:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.731171 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.744857 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.749990 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9b4de769-807e-43ca-b0e6-c3248bb4a6c1-serviceca\") pod \"node-ca-28p27\" (UID: \"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\") " pod="openshift-image-registry/node-ca-28p27" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.750042 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9b4de769-807e-43ca-b0e6-c3248bb4a6c1-host\") pod \"node-ca-28p27\" (UID: \"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\") " pod="openshift-image-registry/node-ca-28p27" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.750087 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prdf5\" (UniqueName: \"kubernetes.io/projected/9b4de769-807e-43ca-b0e6-c3248bb4a6c1-kube-api-access-prdf5\") pod \"node-ca-28p27\" (UID: \"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\") " pod="openshift-image-registry/node-ca-28p27" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.760374 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.773138 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.781499 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.792446 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.805215 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.820563 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.824016 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.824052 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.824063 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.824079 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.824087 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:32Z","lastTransitionTime":"2026-01-21T17:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.832980 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.845920 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.851411 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prdf5\" (UniqueName: \"kubernetes.io/projected/9b4de769-807e-43ca-b0e6-c3248bb4a6c1-kube-api-access-prdf5\") pod \"node-ca-28p27\" (UID: \"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\") " pod="openshift-image-registry/node-ca-28p27" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.851493 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9b4de769-807e-43ca-b0e6-c3248bb4a6c1-serviceca\") pod \"node-ca-28p27\" (UID: \"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\") " pod="openshift-image-registry/node-ca-28p27" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.851530 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9b4de769-807e-43ca-b0e6-c3248bb4a6c1-host\") pod \"node-ca-28p27\" (UID: \"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\") " pod="openshift-image-registry/node-ca-28p27" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.851588 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9b4de769-807e-43ca-b0e6-c3248bb4a6c1-host\") pod \"node-ca-28p27\" (UID: \"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\") " pod="openshift-image-registry/node-ca-28p27" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.853373 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9b4de769-807e-43ca-b0e6-c3248bb4a6c1-serviceca\") pod \"node-ca-28p27\" (UID: \"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\") " pod="openshift-image-registry/node-ca-28p27" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.859376 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.870539 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prdf5\" (UniqueName: \"kubernetes.io/projected/9b4de769-807e-43ca-b0e6-c3248bb4a6c1-kube-api-access-prdf5\") pod \"node-ca-28p27\" (UID: \"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\") " pod="openshift-image-registry/node-ca-28p27" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.871602 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.886184 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.902234 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.913226 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.925801 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.926689 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.926788 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.926858 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.926922 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.926977 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:32Z","lastTransitionTime":"2026-01-21T17:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.937957 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.953466 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.966113 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-28p27" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.967081 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:32 crc kubenswrapper[4808]: I0121 17:56:32.981602 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:32.999920 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:32Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.012641 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.027200 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.029847 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.029880 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.029888 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.029901 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.029911 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:33Z","lastTransitionTime":"2026-01-21T17:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.040403 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.133561 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.133611 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.133623 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.133644 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.133701 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:33Z","lastTransitionTime":"2026-01-21T17:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.237428 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.237477 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.237490 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.237508 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.237520 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:33Z","lastTransitionTime":"2026-01-21T17:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.340565 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.340936 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.340949 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.340968 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.340978 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:33Z","lastTransitionTime":"2026-01-21T17:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.444019 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.444059 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.444073 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.444091 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.444102 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:33Z","lastTransitionTime":"2026-01-21T17:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.510696 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 00:11:06.163096699 +0000 UTC Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.520122 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.520196 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.520139 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:33 crc kubenswrapper[4808]: E0121 17:56:33.520299 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:56:33 crc kubenswrapper[4808]: E0121 17:56:33.520453 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:56:33 crc kubenswrapper[4808]: E0121 17:56:33.520558 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.546676 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.546722 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.546732 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.546751 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.546760 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:33Z","lastTransitionTime":"2026-01-21T17:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.648600 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.648638 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.648649 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.648667 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.648679 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:33Z","lastTransitionTime":"2026-01-21T17:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.674538 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-28p27" event={"ID":"9b4de769-807e-43ca-b0e6-c3248bb4a6c1","Type":"ContainerStarted","Data":"92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521"} Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.674581 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-28p27" event={"ID":"9b4de769-807e-43ca-b0e6-c3248bb4a6c1","Type":"ContainerStarted","Data":"013b607d5ad2e681703fe1c4d7880a95f265a4e4eff5f39854793fa8755dcf5d"} Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.676664 4808 generic.go:334] "Generic (PLEG): container finished" podID="4ffed385-f2d1-4043-9323-561451d85fad" containerID="759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17" exitCode=0 Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.676695 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" event={"ID":"4ffed385-f2d1-4043-9323-561451d85fad","Type":"ContainerDied","Data":"759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17"} Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.687197 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.699598 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.711630 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.722533 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.736120 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.760900 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.760937 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.760946 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.760962 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.760972 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:33Z","lastTransitionTime":"2026-01-21T17:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.781053 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.803270 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.820729 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.840534 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.854315 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.862904 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.862937 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.862947 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.862973 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.862984 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:33Z","lastTransitionTime":"2026-01-21T17:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.867132 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.883857 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.906381 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.921804 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.939801 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.953423 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.965936 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.965980 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.965995 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.966017 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.966030 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:33Z","lastTransitionTime":"2026-01-21T17:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.976270 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.988598 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:33 crc kubenswrapper[4808]: I0121 17:56:33.999712 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:33Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.010359 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.018908 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.030670 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.045463 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.057531 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.068630 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.068668 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.068677 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.068692 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.068701 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:34Z","lastTransitionTime":"2026-01-21T17:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.071975 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.091285 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.107662 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.126085 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.170842 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.170876 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.170884 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.170899 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.170909 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:34Z","lastTransitionTime":"2026-01-21T17:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.273332 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.273368 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.273379 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.273395 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.273405 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:34Z","lastTransitionTime":"2026-01-21T17:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.377717 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.377795 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.377818 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.377847 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.377868 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:34Z","lastTransitionTime":"2026-01-21T17:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.481224 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.481313 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.481335 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.481364 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.481385 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:34Z","lastTransitionTime":"2026-01-21T17:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.511569 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 22:02:12.402634308 +0000 UTC Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.584535 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.584574 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.584584 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.584885 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.584902 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:34Z","lastTransitionTime":"2026-01-21T17:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.688737 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.688784 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.688796 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.688816 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.688829 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:34Z","lastTransitionTime":"2026-01-21T17:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.691300 4808 generic.go:334] "Generic (PLEG): container finished" podID="4ffed385-f2d1-4043-9323-561451d85fad" containerID="3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169" exitCode=0 Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.691380 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" event={"ID":"4ffed385-f2d1-4043-9323-561451d85fad","Type":"ContainerDied","Data":"3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169"} Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.699744 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerStarted","Data":"23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae"} Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.708776 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.723672 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.736962 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.747778 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.761317 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.772733 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.788550 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.792234 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.792289 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.792302 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.792319 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.792331 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:34Z","lastTransitionTime":"2026-01-21T17:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.803965 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.818815 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.835021 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.855487 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.872822 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.890552 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.894160 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.894209 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.894221 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.894274 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.894287 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:34Z","lastTransitionTime":"2026-01-21T17:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.910093 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:34Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.996802 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.996841 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.996854 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.996870 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:34 crc kubenswrapper[4808]: I0121 17:56:34.996880 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:34Z","lastTransitionTime":"2026-01-21T17:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.099908 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.099956 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.099967 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.099984 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.099996 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:35Z","lastTransitionTime":"2026-01-21T17:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.202697 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.202735 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.202747 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.202764 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.202777 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:35Z","lastTransitionTime":"2026-01-21T17:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.304625 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.304668 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.304677 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.304692 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.304701 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:35Z","lastTransitionTime":"2026-01-21T17:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.404109 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.404820 4808 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.419108 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.419158 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.419176 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.419198 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.419276 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:35Z","lastTransitionTime":"2026-01-21T17:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.511933 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 22:00:39.22785516 +0000 UTC Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.520520 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.520734 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.520543 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:35 crc kubenswrapper[4808]: E0121 17:56:35.521080 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:56:35 crc kubenswrapper[4808]: E0121 17:56:35.521235 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:56:35 crc kubenswrapper[4808]: E0121 17:56:35.521401 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.523707 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.523822 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.523904 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.524005 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.524090 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:35Z","lastTransitionTime":"2026-01-21T17:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.626708 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.626773 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.626794 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.626820 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.626837 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:35Z","lastTransitionTime":"2026-01-21T17:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.705202 4808 generic.go:334] "Generic (PLEG): container finished" podID="4ffed385-f2d1-4043-9323-561451d85fad" containerID="b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e" exitCode=0 Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.705286 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" event={"ID":"4ffed385-f2d1-4043-9323-561451d85fad","Type":"ContainerDied","Data":"b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e"} Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.729765 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.729832 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.729848 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.729877 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.729895 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:35Z","lastTransitionTime":"2026-01-21T17:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.832421 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.832462 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.832473 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.832490 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.832501 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:35Z","lastTransitionTime":"2026-01-21T17:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.935742 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.936059 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.936072 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.936088 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:35 crc kubenswrapper[4808]: I0121 17:56:35.936100 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:35Z","lastTransitionTime":"2026-01-21T17:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.039530 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.039585 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.039598 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.039619 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.039633 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:36Z","lastTransitionTime":"2026-01-21T17:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.141863 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.141915 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.141933 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.141958 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.141976 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:36Z","lastTransitionTime":"2026-01-21T17:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.244599 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.244678 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.244704 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.244734 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.244802 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:36Z","lastTransitionTime":"2026-01-21T17:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.347586 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.347658 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.347681 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.347711 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.347734 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:36Z","lastTransitionTime":"2026-01-21T17:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.431188 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.450706 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.450770 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.450787 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.450811 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.450827 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:36Z","lastTransitionTime":"2026-01-21T17:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.452303 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.468740 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.481220 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.496298 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.507958 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.513275 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 20:15:48.088032161 +0000 UTC Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.525711 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.546003 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.552757 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.552788 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.552798 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.552814 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.552825 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:36Z","lastTransitionTime":"2026-01-21T17:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.562594 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.581468 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.597737 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.613037 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.627297 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.649343 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.655468 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.655524 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.655543 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.655570 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.655592 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:36Z","lastTransitionTime":"2026-01-21T17:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.662135 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.677833 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.695055 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.712402 4808 generic.go:334] "Generic (PLEG): container finished" podID="4ffed385-f2d1-4043-9323-561451d85fad" containerID="b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc" exitCode=0 Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.712438 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" event={"ID":"4ffed385-f2d1-4043-9323-561451d85fad","Type":"ContainerDied","Data":"b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc"} Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.715364 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.721653 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerStarted","Data":"92ff75738eebcfa4d50fa93f10b1b83300015364b5d04f18ca764d3da2e514c3"} Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.722549 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.722590 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.735848 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.750447 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.750841 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.757650 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.757683 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.757693 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.757709 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.757725 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:36Z","lastTransitionTime":"2026-01-21T17:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.757953 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.769599 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.783112 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.799643 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.820013 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.831899 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.844827 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.858965 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.860637 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.860681 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.860692 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.860707 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.860717 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:36Z","lastTransitionTime":"2026-01-21T17:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.868956 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.881013 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.891784 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.902378 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.913609 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.923696 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.938635 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.950337 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.960941 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.962459 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.962501 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.962515 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.962531 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.962543 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:36Z","lastTransitionTime":"2026-01-21T17:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.970901 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.982410 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:36 crc kubenswrapper[4808]: I0121 17:56:36.992883 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:36Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.006622 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.021979 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.037281 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.037322 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.037331 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.037347 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.037359 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:37Z","lastTransitionTime":"2026-01-21T17:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.044438 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ff75738eebcfa4d50fa93f10b1b83300015364b5d04f18ca764d3da2e514c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: E0121 17:56:37.054151 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.057780 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.057831 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.057846 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.057863 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.057874 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:37Z","lastTransitionTime":"2026-01-21T17:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.059755 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: E0121 17:56:37.072120 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.075320 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.075713 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.075743 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.075755 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.075770 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.075781 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:37Z","lastTransitionTime":"2026-01-21T17:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.086594 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: E0121 17:56:37.088671 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.091777 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.091816 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.091827 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.091841 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.091851 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:37Z","lastTransitionTime":"2026-01-21T17:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.097198 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: E0121 17:56:37.103329 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.107920 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.107970 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.107982 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.108003 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.108015 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:37Z","lastTransitionTime":"2026-01-21T17:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.110636 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: E0121 17:56:37.118973 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: E0121 17:56:37.119107 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.120477 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.120525 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.120571 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.120595 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.120610 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:37Z","lastTransitionTime":"2026-01-21T17:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.124671 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.136094 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.150542 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.165654 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.178764 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.189834 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.204658 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.220527 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.222342 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.222372 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.222383 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.222397 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.222407 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:37Z","lastTransitionTime":"2026-01-21T17:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.237215 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ff75738eebcfa4d50fa93f10b1b83300015364b5d04f18ca764d3da2e514c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.325184 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.325275 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.325294 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.325320 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.325338 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:37Z","lastTransitionTime":"2026-01-21T17:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.427780 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.427847 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.427865 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.427893 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.427912 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:37Z","lastTransitionTime":"2026-01-21T17:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.513884 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 23:42:49.848726244 +0000 UTC Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.520235 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.520332 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:37 crc kubenswrapper[4808]: E0121 17:56:37.520411 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.520449 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:37 crc kubenswrapper[4808]: E0121 17:56:37.520667 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:56:37 crc kubenswrapper[4808]: E0121 17:56:37.520918 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.531174 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.531224 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.531240 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.531278 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.531290 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:37Z","lastTransitionTime":"2026-01-21T17:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.633977 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.634057 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.634081 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.634115 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.634138 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:37Z","lastTransitionTime":"2026-01-21T17:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.734582 4808 generic.go:334] "Generic (PLEG): container finished" podID="4ffed385-f2d1-4043-9323-561451d85fad" containerID="dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84" exitCode=0 Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.734690 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" event={"ID":"4ffed385-f2d1-4043-9323-561451d85fad","Type":"ContainerDied","Data":"dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84"} Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.734829 4808 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.742124 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.742394 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.742410 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.742431 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.742451 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:37Z","lastTransitionTime":"2026-01-21T17:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.757114 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.771158 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.786293 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.802650 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.823657 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.844418 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.846627 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.846720 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.846739 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.846765 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.846781 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:37Z","lastTransitionTime":"2026-01-21T17:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.863559 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.884818 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.907363 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.925611 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.936632 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.949760 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.949838 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.949855 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.949879 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.949894 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:37Z","lastTransitionTime":"2026-01-21T17:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.956767 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:37 crc kubenswrapper[4808]: I0121 17:56:37.983994 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ff75738eebcfa4d50fa93f10b1b83300015364b5d04f18ca764d3da2e514c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.001959 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:37Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.053694 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.053770 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.053789 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.053809 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.053855 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:38Z","lastTransitionTime":"2026-01-21T17:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.177910 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.177964 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.177979 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.178002 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.178031 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:38Z","lastTransitionTime":"2026-01-21T17:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.280562 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.280596 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.280604 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.280619 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.280628 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:38Z","lastTransitionTime":"2026-01-21T17:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.383006 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.383045 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.383054 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.383068 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.383077 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:38Z","lastTransitionTime":"2026-01-21T17:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.485998 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.486036 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.486046 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.486062 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.486075 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:38Z","lastTransitionTime":"2026-01-21T17:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.515330 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 12:36:36.714749976 +0000 UTC Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.588710 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.588758 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.588772 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.588788 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.588799 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:38Z","lastTransitionTime":"2026-01-21T17:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.691434 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.691767 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.691776 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.691791 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.691800 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:38Z","lastTransitionTime":"2026-01-21T17:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.743096 4808 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.744205 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" event={"ID":"4ffed385-f2d1-4043-9323-561451d85fad","Type":"ContainerStarted","Data":"710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02"} Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.756888 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:38Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.774540 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:38Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.793732 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:38Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.794755 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.794801 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.794812 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.794831 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.794844 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:38Z","lastTransitionTime":"2026-01-21T17:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.808098 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:38Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.859357 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:38Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.877638 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:38Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.892915 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:38Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.897788 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.897826 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.897837 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.897856 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.897867 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:38Z","lastTransitionTime":"2026-01-21T17:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.910206 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:38Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.924085 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:38Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.941922 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ff75738eebcfa4d50fa93f10b1b83300015364b5d04f18ca764d3da2e514c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:38Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.955710 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:38Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.966827 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:38Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.979605 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:38Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:38 crc kubenswrapper[4808]: I0121 17:56:38.991619 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:38Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.000322 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.000366 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.000376 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.000393 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.000405 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:39Z","lastTransitionTime":"2026-01-21T17:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.102900 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.102956 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.102968 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.102988 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.103003 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:39Z","lastTransitionTime":"2026-01-21T17:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.206408 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.206456 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.206473 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.206496 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.206512 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:39Z","lastTransitionTime":"2026-01-21T17:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.309609 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.309675 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.309696 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.309721 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.309737 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:39Z","lastTransitionTime":"2026-01-21T17:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.320199 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:56:39 crc kubenswrapper[4808]: E0121 17:56:39.320460 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:56:55.320428051 +0000 UTC m=+50.050938946 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.413702 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.413770 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.413792 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.413824 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.413842 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:39Z","lastTransitionTime":"2026-01-21T17:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.421464 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.421546 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.421598 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.421643 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:39 crc kubenswrapper[4808]: E0121 17:56:39.421706 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 17:56:39 crc kubenswrapper[4808]: E0121 17:56:39.421801 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 17:56:39 crc kubenswrapper[4808]: E0121 17:56:39.421828 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:55.42179813 +0000 UTC m=+50.152309055 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 17:56:39 crc kubenswrapper[4808]: E0121 17:56:39.421850 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 17:56:39 crc kubenswrapper[4808]: E0121 17:56:39.421889 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:55.421866262 +0000 UTC m=+50.152377187 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 17:56:39 crc kubenswrapper[4808]: E0121 17:56:39.421897 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 17:56:39 crc kubenswrapper[4808]: E0121 17:56:39.421926 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:39 crc kubenswrapper[4808]: E0121 17:56:39.421922 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 17:56:39 crc kubenswrapper[4808]: E0121 17:56:39.421980 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 17:56:39 crc kubenswrapper[4808]: E0121 17:56:39.422012 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:39 crc kubenswrapper[4808]: E0121 17:56:39.422018 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:55.421992084 +0000 UTC m=+50.152503029 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:39 crc kubenswrapper[4808]: E0121 17:56:39.422115 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:55.422089286 +0000 UTC m=+50.152600171 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.515776 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 03:55:59.077798752 +0000 UTC Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.517662 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.517711 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.517724 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.517745 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.517760 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:39Z","lastTransitionTime":"2026-01-21T17:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.520190 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.520190 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.520279 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:39 crc kubenswrapper[4808]: E0121 17:56:39.520348 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:56:39 crc kubenswrapper[4808]: E0121 17:56:39.520470 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:56:39 crc kubenswrapper[4808]: E0121 17:56:39.520542 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.621066 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.621124 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.621142 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.621170 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.621190 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:39Z","lastTransitionTime":"2026-01-21T17:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.724369 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.724427 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.724440 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.724461 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.724476 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:39Z","lastTransitionTime":"2026-01-21T17:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.750132 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovnkube-controller/0.log" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.753418 4808 generic.go:334] "Generic (PLEG): container finished" podID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerID="92ff75738eebcfa4d50fa93f10b1b83300015364b5d04f18ca764d3da2e514c3" exitCode=1 Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.753474 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerDied","Data":"92ff75738eebcfa4d50fa93f10b1b83300015364b5d04f18ca764d3da2e514c3"} Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.754355 4808 scope.go:117] "RemoveContainer" containerID="92ff75738eebcfa4d50fa93f10b1b83300015364b5d04f18ca764d3da2e514c3" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.772003 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:39Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.784555 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:39Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.796292 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:39Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.810366 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:39Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.821777 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:39Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.826150 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.826176 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.826185 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.826202 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.826212 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:39Z","lastTransitionTime":"2026-01-21T17:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.834763 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:39Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.852154 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:39Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.869989 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:39Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.886655 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:39Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.906822 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ff75738eebcfa4d50fa93f10b1b83300015364b5d04f18ca764d3da2e514c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ff75738eebcfa4d50fa93f10b1b83300015364b5d04f18ca764d3da2e514c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:39Z\\\",\\\"message\\\":\\\"emoval\\\\nI0121 17:56:38.893906 6079 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0121 17:56:38.893923 6079 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0121 17:56:38.893929 6079 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0121 17:56:38.893948 6079 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0121 17:56:38.893956 6079 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0121 17:56:38.893965 6079 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0121 17:56:38.893990 6079 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0121 17:56:38.894015 6079 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0121 17:56:38.894025 6079 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0121 17:56:38.894028 6079 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0121 17:56:38.894051 6079 factory.go:656] Stopping watch factory\\\\nI0121 17:56:38.894054 6079 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0121 17:56:38.894069 6079 ovnkube.go:599] Stopped ovnkube\\\\nI0121 17:56:38.894082 6079 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0121 17:56:38.894101 6079 handler.go:208] Removed *v1.Node event handler 7\\\\nI0121 17:56:38.894102 6079 metrics.go:553] Stopping metrics server at address \\\\\\\"127.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:39Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.922819 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:39Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.929328 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.929366 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.929375 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.929391 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.929403 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:39Z","lastTransitionTime":"2026-01-21T17:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.935936 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:39Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.954658 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:39Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:39 crc kubenswrapper[4808]: I0121 17:56:39.969450 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:39Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.033084 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.033134 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.033149 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.033171 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.033184 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:40Z","lastTransitionTime":"2026-01-21T17:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.136094 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.136135 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.136146 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.136162 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.136173 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:40Z","lastTransitionTime":"2026-01-21T17:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.146255 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.238574 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.238624 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.238636 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.238654 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.238666 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:40Z","lastTransitionTime":"2026-01-21T17:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.341273 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.341312 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.341322 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.341337 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.341346 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:40Z","lastTransitionTime":"2026-01-21T17:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.443667 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.443715 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.443724 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.443739 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.443750 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:40Z","lastTransitionTime":"2026-01-21T17:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.516382 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 17:40:40.546057728 +0000 UTC Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.545435 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.545476 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.545487 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.545501 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.545512 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:40Z","lastTransitionTime":"2026-01-21T17:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.648685 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.648743 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.648756 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.648776 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.648789 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:40Z","lastTransitionTime":"2026-01-21T17:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.752165 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.752225 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.752291 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.752325 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.752346 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:40Z","lastTransitionTime":"2026-01-21T17:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.758317 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovnkube-controller/0.log" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.761472 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerStarted","Data":"019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0"} Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.762065 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.777395 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:40Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.793820 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:40Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.808082 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:40Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.819009 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:40Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.831867 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:40Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.849843 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:40Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.854499 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.854591 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.854603 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.854618 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.854628 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:40Z","lastTransitionTime":"2026-01-21T17:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.865113 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:40Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.880412 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:40Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.893856 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:40Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.905556 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:40Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.927688 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ff75738eebcfa4d50fa93f10b1b83300015364b5d04f18ca764d3da2e514c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:39Z\\\",\\\"message\\\":\\\"emoval\\\\nI0121 17:56:38.893906 6079 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0121 17:56:38.893923 6079 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0121 17:56:38.893929 6079 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0121 17:56:38.893948 6079 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0121 17:56:38.893956 6079 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0121 17:56:38.893965 6079 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0121 17:56:38.893990 6079 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0121 17:56:38.894015 6079 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0121 17:56:38.894025 6079 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0121 17:56:38.894028 6079 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0121 17:56:38.894051 6079 factory.go:656] Stopping watch factory\\\\nI0121 17:56:38.894054 6079 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0121 17:56:38.894069 6079 ovnkube.go:599] Stopped ovnkube\\\\nI0121 17:56:38.894082 6079 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0121 17:56:38.894101 6079 handler.go:208] Removed *v1.Node event handler 7\\\\nI0121 17:56:38.894102 6079 metrics.go:553] Stopping metrics server at address \\\\\\\"127.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:40Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.946784 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:40Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.957739 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.957798 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.957822 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.957852 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.957875 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:40Z","lastTransitionTime":"2026-01-21T17:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.961491 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:40Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:40 crc kubenswrapper[4808]: I0121 17:56:40.972385 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:40Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.061115 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.061172 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.061191 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.061215 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.061232 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:41Z","lastTransitionTime":"2026-01-21T17:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.164215 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.164327 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.164346 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.164376 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.164394 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:41Z","lastTransitionTime":"2026-01-21T17:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.267182 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.267221 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.267232 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.267268 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.267280 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:41Z","lastTransitionTime":"2026-01-21T17:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.370151 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.370216 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.370231 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.370270 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.370285 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:41Z","lastTransitionTime":"2026-01-21T17:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.473808 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.473879 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.473900 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.473933 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.473962 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:41Z","lastTransitionTime":"2026-01-21T17:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.516544 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 18:25:47.196662579 +0000 UTC Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.520072 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:41 crc kubenswrapper[4808]: E0121 17:56:41.520360 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.520609 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.520722 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:41 crc kubenswrapper[4808]: E0121 17:56:41.520817 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:56:41 crc kubenswrapper[4808]: E0121 17:56:41.520955 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.577858 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.577951 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.577980 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.578017 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.578049 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:41Z","lastTransitionTime":"2026-01-21T17:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.681234 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.681318 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.681328 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.681348 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.681361 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:41Z","lastTransitionTime":"2026-01-21T17:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.768744 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovnkube-controller/1.log" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.770062 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovnkube-controller/0.log" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.773802 4808 generic.go:334] "Generic (PLEG): container finished" podID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerID="019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0" exitCode=1 Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.773860 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerDied","Data":"019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0"} Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.773955 4808 scope.go:117] "RemoveContainer" containerID="92ff75738eebcfa4d50fa93f10b1b83300015364b5d04f18ca764d3da2e514c3" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.775541 4808 scope.go:117] "RemoveContainer" containerID="019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0" Jan 21 17:56:41 crc kubenswrapper[4808]: E0121 17:56:41.776086 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fc2fk_openshift-ovn-kubernetes(bb13a20f-f4ab-412c-8165-5b93edc79628)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.783788 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.783833 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.783847 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.783867 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.783884 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:41Z","lastTransitionTime":"2026-01-21T17:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.794971 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:41Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.816655 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:41Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.832580 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:41Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.850648 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:41Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.867823 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:41Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.884186 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:41Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.887339 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.887374 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.887386 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.887404 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.887421 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:41Z","lastTransitionTime":"2026-01-21T17:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.901223 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:41Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.923442 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:41Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.940792 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:41Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.955644 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:41Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.981796 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g"] Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.982471 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.983701 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ff75738eebcfa4d50fa93f10b1b83300015364b5d04f18ca764d3da2e514c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:39Z\\\",\\\"message\\\":\\\"emoval\\\\nI0121 17:56:38.893906 6079 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0121 17:56:38.893923 6079 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0121 17:56:38.893929 6079 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0121 17:56:38.893948 6079 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0121 17:56:38.893956 6079 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0121 17:56:38.893965 6079 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0121 17:56:38.893990 6079 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0121 17:56:38.894015 6079 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0121 17:56:38.894025 6079 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0121 17:56:38.894028 6079 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0121 17:56:38.894051 6079 factory.go:656] Stopping watch factory\\\\nI0121 17:56:38.894054 6079 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0121 17:56:38.894069 6079 ovnkube.go:599] Stopped ovnkube\\\\nI0121 17:56:38.894082 6079 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0121 17:56:38.894101 6079 handler.go:208] Removed *v1.Node event handler 7\\\\nI0121 17:56:38.894102 6079 metrics.go:553] Stopping metrics server at address \\\\\\\"127.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:40Z\\\",\\\"message\\\":\\\"ectMeta:{metrics openshift-etcd-operator ff1da138-ae82-4792-ae1f-3b2df1427723 4289 0 2025-02-23 05:12:19 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:etcd-operator] map[include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-secret-name:etcd-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00078f0a7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: etcd-operator,},ClusterIP:10.217.5.188,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.188],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Conditi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:41Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.985076 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.989693 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.992430 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.992480 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.992495 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.992517 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:41 crc kubenswrapper[4808]: I0121 17:56:41.992530 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:41Z","lastTransitionTime":"2026-01-21T17:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.003167 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.017286 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.029990 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.041788 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.051824 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.052261 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/53885bfd-41c1-49a4-beb3-e937534a6ef8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mw46g\" (UID: \"53885bfd-41c1-49a4-beb3-e937534a6ef8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.052351 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/53885bfd-41c1-49a4-beb3-e937534a6ef8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mw46g\" (UID: \"53885bfd-41c1-49a4-beb3-e937534a6ef8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.052386 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/53885bfd-41c1-49a4-beb3-e937534a6ef8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mw46g\" (UID: \"53885bfd-41c1-49a4-beb3-e937534a6ef8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.052630 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tn7ts\" (UniqueName: \"kubernetes.io/projected/53885bfd-41c1-49a4-beb3-e937534a6ef8-kube-api-access-tn7ts\") pod \"ovnkube-control-plane-749d76644c-mw46g\" (UID: \"53885bfd-41c1-49a4-beb3-e937534a6ef8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.064702 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.076221 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.092196 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53885bfd-41c1-49a4-beb3-e937534a6ef8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw46g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.095162 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.095203 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.095214 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.095232 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.095261 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:42Z","lastTransitionTime":"2026-01-21T17:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.105687 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.120722 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.140534 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.154061 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tn7ts\" (UniqueName: \"kubernetes.io/projected/53885bfd-41c1-49a4-beb3-e937534a6ef8-kube-api-access-tn7ts\") pod \"ovnkube-control-plane-749d76644c-mw46g\" (UID: \"53885bfd-41c1-49a4-beb3-e937534a6ef8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.154105 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/53885bfd-41c1-49a4-beb3-e937534a6ef8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mw46g\" (UID: \"53885bfd-41c1-49a4-beb3-e937534a6ef8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.154168 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/53885bfd-41c1-49a4-beb3-e937534a6ef8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mw46g\" (UID: \"53885bfd-41c1-49a4-beb3-e937534a6ef8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.154225 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/53885bfd-41c1-49a4-beb3-e937534a6ef8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mw46g\" (UID: \"53885bfd-41c1-49a4-beb3-e937534a6ef8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.154882 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/53885bfd-41c1-49a4-beb3-e937534a6ef8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mw46g\" (UID: \"53885bfd-41c1-49a4-beb3-e937534a6ef8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.157157 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/53885bfd-41c1-49a4-beb3-e937534a6ef8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mw46g\" (UID: \"53885bfd-41c1-49a4-beb3-e937534a6ef8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.159443 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.167383 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/53885bfd-41c1-49a4-beb3-e937534a6ef8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mw46g\" (UID: \"53885bfd-41c1-49a4-beb3-e937534a6ef8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.179169 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tn7ts\" (UniqueName: \"kubernetes.io/projected/53885bfd-41c1-49a4-beb3-e937534a6ef8-kube-api-access-tn7ts\") pod \"ovnkube-control-plane-749d76644c-mw46g\" (UID: \"53885bfd-41c1-49a4-beb3-e937534a6ef8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.188779 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.198030 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.198102 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.198121 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.198148 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.198169 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:42Z","lastTransitionTime":"2026-01-21T17:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.207791 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.222473 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.240677 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.266286 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ff75738eebcfa4d50fa93f10b1b83300015364b5d04f18ca764d3da2e514c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:39Z\\\",\\\"message\\\":\\\"emoval\\\\nI0121 17:56:38.893906 6079 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0121 17:56:38.893923 6079 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0121 17:56:38.893929 6079 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0121 17:56:38.893948 6079 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0121 17:56:38.893956 6079 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0121 17:56:38.893965 6079 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0121 17:56:38.893990 6079 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0121 17:56:38.894015 6079 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0121 17:56:38.894025 6079 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0121 17:56:38.894028 6079 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0121 17:56:38.894051 6079 factory.go:656] Stopping watch factory\\\\nI0121 17:56:38.894054 6079 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0121 17:56:38.894069 6079 ovnkube.go:599] Stopped ovnkube\\\\nI0121 17:56:38.894082 6079 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0121 17:56:38.894101 6079 handler.go:208] Removed *v1.Node event handler 7\\\\nI0121 17:56:38.894102 6079 metrics.go:553] Stopping metrics server at address \\\\\\\"127.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:40Z\\\",\\\"message\\\":\\\"ectMeta:{metrics openshift-etcd-operator ff1da138-ae82-4792-ae1f-3b2df1427723 4289 0 2025-02-23 05:12:19 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:etcd-operator] map[include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-secret-name:etcd-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00078f0a7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: etcd-operator,},ClusterIP:10.217.5.188,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.188],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Conditi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.291173 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.301174 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.301895 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.301936 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.301949 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.301966 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.301980 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:42Z","lastTransitionTime":"2026-01-21T17:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:42 crc kubenswrapper[4808]: W0121 17:56:42.320328 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53885bfd_41c1_49a4_beb3_e937534a6ef8.slice/crio-7fa5ba1d541d12e5d976c1503ba0a855e767f9a5d9572bcb253740e0e1ba059d WatchSource:0}: Error finding container 7fa5ba1d541d12e5d976c1503ba0a855e767f9a5d9572bcb253740e0e1ba059d: Status 404 returned error can't find the container with id 7fa5ba1d541d12e5d976c1503ba0a855e767f9a5d9572bcb253740e0e1ba059d Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.405317 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.405358 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.405369 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.405388 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.405402 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:42Z","lastTransitionTime":"2026-01-21T17:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.507407 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.507720 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.507728 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.507741 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.507750 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:42Z","lastTransitionTime":"2026-01-21T17:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.517452 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 20:59:24.564857943 +0000 UTC Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.610312 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.610364 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.610375 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.610395 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.610407 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:42Z","lastTransitionTime":"2026-01-21T17:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.712981 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.713029 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.713040 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.713058 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.713068 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:42Z","lastTransitionTime":"2026-01-21T17:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.777887 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovnkube-controller/1.log" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.781254 4808 scope.go:117] "RemoveContainer" containerID="019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0" Jan 21 17:56:42 crc kubenswrapper[4808]: E0121 17:56:42.781455 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fc2fk_openshift-ovn-kubernetes(bb13a20f-f4ab-412c-8165-5b93edc79628)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.781883 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" event={"ID":"53885bfd-41c1-49a4-beb3-e937534a6ef8","Type":"ContainerStarted","Data":"90dfe0143cdd3cd582497eddee45eeb42dd96a03eb26a3b753c8aa2fb37436bd"} Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.781921 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" event={"ID":"53885bfd-41c1-49a4-beb3-e937534a6ef8","Type":"ContainerStarted","Data":"3ce09f9763152c19c6f0cc141317c798d49f50066d1f0184a114085a19fc9adf"} Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.781932 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" event={"ID":"53885bfd-41c1-49a4-beb3-e937534a6ef8","Type":"ContainerStarted","Data":"7fa5ba1d541d12e5d976c1503ba0a855e767f9a5d9572bcb253740e0e1ba059d"} Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.792937 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.808408 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53885bfd-41c1-49a4-beb3-e937534a6ef8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw46g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.815547 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.815593 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.815609 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.815629 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.815642 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:42Z","lastTransitionTime":"2026-01-21T17:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.819813 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.831319 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.844272 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.859167 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.869575 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.885042 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.900491 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.918461 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.918526 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.918540 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.918562 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.918577 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:42Z","lastTransitionTime":"2026-01-21T17:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.918977 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:40Z\\\",\\\"message\\\":\\\"ectMeta:{metrics openshift-etcd-operator ff1da138-ae82-4792-ae1f-3b2df1427723 4289 0 2025-02-23 05:12:19 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:etcd-operator] map[include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-secret-name:etcd-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00078f0a7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: etcd-operator,},ClusterIP:10.217.5.188,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.188],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Conditi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fc2fk_openshift-ovn-kubernetes(bb13a20f-f4ab-412c-8165-5b93edc79628)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.933395 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.945394 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.959415 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.976354 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:42 crc kubenswrapper[4808]: I0121 17:56:42.989680 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:42Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.003001 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.017972 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.020738 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.020794 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.020808 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.020827 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.020839 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:43Z","lastTransitionTime":"2026-01-21T17:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.033369 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.045407 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.057957 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.072559 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.086334 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.104225 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.118618 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.122812 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.122863 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.122937 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.122962 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.123007 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:43Z","lastTransitionTime":"2026-01-21T17:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.137270 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:40Z\\\",\\\"message\\\":\\\"ectMeta:{metrics openshift-etcd-operator ff1da138-ae82-4792-ae1f-3b2df1427723 4289 0 2025-02-23 05:12:19 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:etcd-operator] map[include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-secret-name:etcd-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00078f0a7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: etcd-operator,},ClusterIP:10.217.5.188,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.188],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Conditi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fc2fk_openshift-ovn-kubernetes(bb13a20f-f4ab-412c-8165-5b93edc79628)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.149182 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.162204 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.178392 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.189707 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.200364 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53885bfd-41c1-49a4-beb3-e937534a6ef8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce09f9763152c19c6f0cc141317c798d49f50066d1f0184a114085a19fc9adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90dfe0143cdd3cd582497eddee45eeb42dd96a03eb26a3b753c8aa2fb37436bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw46g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.225527 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.225575 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.225589 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.225611 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.225626 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:43Z","lastTransitionTime":"2026-01-21T17:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.328295 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.328350 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.328366 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.328390 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.328408 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:43Z","lastTransitionTime":"2026-01-21T17:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.431223 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.431279 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.431297 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.431323 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.431338 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:43Z","lastTransitionTime":"2026-01-21T17:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.466014 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-gm2t2"] Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.466686 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:56:43 crc kubenswrapper[4808]: E0121 17:56:43.466776 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.492147 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.511057 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.518230 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 02:45:37.937899236 +0000 UTC Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.520592 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.520612 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:43 crc kubenswrapper[4808]: E0121 17:56:43.520834 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:56:43 crc kubenswrapper[4808]: E0121 17:56:43.521026 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.521549 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:43 crc kubenswrapper[4808]: E0121 17:56:43.521719 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.534745 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.534802 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.534820 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.534845 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.534864 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:43Z","lastTransitionTime":"2026-01-21T17:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.539685 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:40Z\\\",\\\"message\\\":\\\"ectMeta:{metrics openshift-etcd-operator ff1da138-ae82-4792-ae1f-3b2df1427723 4289 0 2025-02-23 05:12:19 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:etcd-operator] map[include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-secret-name:etcd-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00078f0a7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: etcd-operator,},ClusterIP:10.217.5.188,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.188],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Conditi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fc2fk_openshift-ovn-kubernetes(bb13a20f-f4ab-412c-8165-5b93edc79628)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.557035 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gm2t2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d39c565-f404-451e-896f-ecc5af76a000\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gm2t2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.570764 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6mrn\" (UniqueName: \"kubernetes.io/projected/8d39c565-f404-451e-896f-ecc5af76a000-kube-api-access-h6mrn\") pod \"network-metrics-daemon-gm2t2\" (UID: \"8d39c565-f404-451e-896f-ecc5af76a000\") " pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.570955 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs\") pod \"network-metrics-daemon-gm2t2\" (UID: \"8d39c565-f404-451e-896f-ecc5af76a000\") " pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.576044 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.593629 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.607871 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.621922 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.635553 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53885bfd-41c1-49a4-beb3-e937534a6ef8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce09f9763152c19c6f0cc141317c798d49f50066d1f0184a114085a19fc9adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90dfe0143cdd3cd582497eddee45eeb42dd96a03eb26a3b753c8aa2fb37436bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw46g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.637931 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.637985 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.637996 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.638021 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.638037 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:43Z","lastTransitionTime":"2026-01-21T17:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.655341 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.672356 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.672881 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6mrn\" (UniqueName: \"kubernetes.io/projected/8d39c565-f404-451e-896f-ecc5af76a000-kube-api-access-h6mrn\") pod \"network-metrics-daemon-gm2t2\" (UID: \"8d39c565-f404-451e-896f-ecc5af76a000\") " pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.672970 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs\") pod \"network-metrics-daemon-gm2t2\" (UID: \"8d39c565-f404-451e-896f-ecc5af76a000\") " pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:56:43 crc kubenswrapper[4808]: E0121 17:56:43.673147 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 17:56:43 crc kubenswrapper[4808]: E0121 17:56:43.673233 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs podName:8d39c565-f404-451e-896f-ecc5af76a000 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:44.173213307 +0000 UTC m=+38.903724202 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs") pod "network-metrics-daemon-gm2t2" (UID: "8d39c565-f404-451e-896f-ecc5af76a000") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.693800 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6mrn\" (UniqueName: \"kubernetes.io/projected/8d39c565-f404-451e-896f-ecc5af76a000-kube-api-access-h6mrn\") pod \"network-metrics-daemon-gm2t2\" (UID: \"8d39c565-f404-451e-896f-ecc5af76a000\") " pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.694853 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.719785 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.740988 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.741059 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.741070 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.741090 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.741107 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:43Z","lastTransitionTime":"2026-01-21T17:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.742736 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.760235 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.780599 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:43Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.844099 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.844193 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.844218 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.844302 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.844331 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:43Z","lastTransitionTime":"2026-01-21T17:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.946906 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.946958 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.946976 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.946997 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:43 crc kubenswrapper[4808]: I0121 17:56:43.947014 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:43Z","lastTransitionTime":"2026-01-21T17:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.049756 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.049795 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.049804 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.049819 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.049829 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:44Z","lastTransitionTime":"2026-01-21T17:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.152480 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.152536 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.152548 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.152596 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.152610 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:44Z","lastTransitionTime":"2026-01-21T17:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.179154 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs\") pod \"network-metrics-daemon-gm2t2\" (UID: \"8d39c565-f404-451e-896f-ecc5af76a000\") " pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:56:44 crc kubenswrapper[4808]: E0121 17:56:44.179318 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 17:56:44 crc kubenswrapper[4808]: E0121 17:56:44.179373 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs podName:8d39c565-f404-451e-896f-ecc5af76a000 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:45.179360769 +0000 UTC m=+39.909871654 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs") pod "network-metrics-daemon-gm2t2" (UID: "8d39c565-f404-451e-896f-ecc5af76a000") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.255153 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.255205 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.255215 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.255231 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.255258 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:44Z","lastTransitionTime":"2026-01-21T17:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.357571 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.357609 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.357618 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.357633 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.357641 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:44Z","lastTransitionTime":"2026-01-21T17:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.459903 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.459956 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.459973 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.459996 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.460013 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:44Z","lastTransitionTime":"2026-01-21T17:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.518642 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 01:42:59.231297327 +0000 UTC Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.562701 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.562764 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.562779 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.562805 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.562822 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:44Z","lastTransitionTime":"2026-01-21T17:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.666318 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.666360 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.666372 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.666390 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.666399 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:44Z","lastTransitionTime":"2026-01-21T17:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.769503 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.769588 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.769606 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.769630 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.769649 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:44Z","lastTransitionTime":"2026-01-21T17:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.873004 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.873119 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.873171 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.873193 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.873208 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:44Z","lastTransitionTime":"2026-01-21T17:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.976856 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.976923 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.976940 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.976964 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:44 crc kubenswrapper[4808]: I0121 17:56:44.976980 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:44Z","lastTransitionTime":"2026-01-21T17:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.079868 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.079907 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.079915 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.079929 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.079939 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:45Z","lastTransitionTime":"2026-01-21T17:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.182730 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.182803 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.182819 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.182841 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.182860 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:45Z","lastTransitionTime":"2026-01-21T17:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.187490 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs\") pod \"network-metrics-daemon-gm2t2\" (UID: \"8d39c565-f404-451e-896f-ecc5af76a000\") " pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:56:45 crc kubenswrapper[4808]: E0121 17:56:45.187748 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 17:56:45 crc kubenswrapper[4808]: E0121 17:56:45.187861 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs podName:8d39c565-f404-451e-896f-ecc5af76a000 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:47.187838821 +0000 UTC m=+41.918349746 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs") pod "network-metrics-daemon-gm2t2" (UID: "8d39c565-f404-451e-896f-ecc5af76a000") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.286006 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.286592 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.286610 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.286626 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.286638 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:45Z","lastTransitionTime":"2026-01-21T17:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.389699 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.389783 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.389812 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.389847 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.389878 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:45Z","lastTransitionTime":"2026-01-21T17:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.492603 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.492677 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.492695 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.492720 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.492737 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:45Z","lastTransitionTime":"2026-01-21T17:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.519884 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.520073 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:45 crc kubenswrapper[4808]: E0121 17:56:45.520285 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.520353 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:45 crc kubenswrapper[4808]: E0121 17:56:45.520428 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.520475 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:56:45 crc kubenswrapper[4808]: E0121 17:56:45.520552 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:56:45 crc kubenswrapper[4808]: E0121 17:56:45.520679 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.520090 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 09:26:28.76585808 +0000 UTC Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.544292 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:45Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.560181 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:45Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.580813 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:45Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.597575 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.597620 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.597634 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.597651 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.597671 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:45Z","lastTransitionTime":"2026-01-21T17:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.603327 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:45Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.620264 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:45Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.632463 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:45Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.648487 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:45Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.667591 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:45Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.690111 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:40Z\\\",\\\"message\\\":\\\"ectMeta:{metrics openshift-etcd-operator ff1da138-ae82-4792-ae1f-3b2df1427723 4289 0 2025-02-23 05:12:19 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:etcd-operator] map[include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-secret-name:etcd-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00078f0a7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: etcd-operator,},ClusterIP:10.217.5.188,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.188],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Conditi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fc2fk_openshift-ovn-kubernetes(bb13a20f-f4ab-412c-8165-5b93edc79628)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:45Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.701829 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.701864 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.701872 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.701890 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.701904 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:45Z","lastTransitionTime":"2026-01-21T17:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.704180 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gm2t2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d39c565-f404-451e-896f-ecc5af76a000\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gm2t2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:45Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.719015 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:45Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.731981 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:45Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.749713 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:45Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.767342 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:45Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.778383 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53885bfd-41c1-49a4-beb3-e937534a6ef8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce09f9763152c19c6f0cc141317c798d49f50066d1f0184a114085a19fc9adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90dfe0143cdd3cd582497eddee45eeb42dd96a03eb26a3b753c8aa2fb37436bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw46g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:45Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.786737 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:45Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.804599 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.804640 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.804652 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.804672 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.804684 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:45Z","lastTransitionTime":"2026-01-21T17:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.908275 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.908334 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.908350 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.908372 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:45 crc kubenswrapper[4808]: I0121 17:56:45.908389 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:45Z","lastTransitionTime":"2026-01-21T17:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.010801 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.010861 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.010890 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.010941 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.010970 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:46Z","lastTransitionTime":"2026-01-21T17:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.114012 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.114071 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.114088 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.114112 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.114130 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:46Z","lastTransitionTime":"2026-01-21T17:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.217467 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.217530 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.217555 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.217585 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.217604 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:46Z","lastTransitionTime":"2026-01-21T17:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.321391 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.321473 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.321491 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.321516 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.321533 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:46Z","lastTransitionTime":"2026-01-21T17:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.424641 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.424686 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.424695 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.424710 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.424718 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:46Z","lastTransitionTime":"2026-01-21T17:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.521533 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 15:50:36.714906337 +0000 UTC Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.528858 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.528926 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.528946 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.528968 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.528992 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:46Z","lastTransitionTime":"2026-01-21T17:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.632665 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.632730 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.632747 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.632772 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.632792 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:46Z","lastTransitionTime":"2026-01-21T17:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.736150 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.736192 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.736201 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.736221 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.736232 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:46Z","lastTransitionTime":"2026-01-21T17:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.839886 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.839991 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.840017 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.840049 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.840069 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:46Z","lastTransitionTime":"2026-01-21T17:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.943351 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.943448 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.943481 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.943518 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:46 crc kubenswrapper[4808]: I0121 17:56:46.943540 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:46Z","lastTransitionTime":"2026-01-21T17:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.047672 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.047743 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.047760 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.047792 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.047809 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:47Z","lastTransitionTime":"2026-01-21T17:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.150647 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.150694 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.150707 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.150726 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.150739 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:47Z","lastTransitionTime":"2026-01-21T17:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.211761 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs\") pod \"network-metrics-daemon-gm2t2\" (UID: \"8d39c565-f404-451e-896f-ecc5af76a000\") " pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:56:47 crc kubenswrapper[4808]: E0121 17:56:47.211930 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 17:56:47 crc kubenswrapper[4808]: E0121 17:56:47.211994 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs podName:8d39c565-f404-451e-896f-ecc5af76a000 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:51.21197761 +0000 UTC m=+45.942488505 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs") pod "network-metrics-daemon-gm2t2" (UID: "8d39c565-f404-451e-896f-ecc5af76a000") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.253657 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.253717 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.253734 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.253759 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.253777 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:47Z","lastTransitionTime":"2026-01-21T17:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.356666 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.356717 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.356729 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.356747 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.356760 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:47Z","lastTransitionTime":"2026-01-21T17:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.409056 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.409116 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.409131 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.409153 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.409167 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:47Z","lastTransitionTime":"2026-01-21T17:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:47 crc kubenswrapper[4808]: E0121 17:56:47.427389 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:47Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.432680 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.432750 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.432775 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.432807 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.432829 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:47Z","lastTransitionTime":"2026-01-21T17:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:47 crc kubenswrapper[4808]: E0121 17:56:47.454925 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:47Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.460168 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.460214 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.460230 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.460268 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.460284 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:47Z","lastTransitionTime":"2026-01-21T17:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:47 crc kubenswrapper[4808]: E0121 17:56:47.487131 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:47Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.492931 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.492990 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.493010 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.493039 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.493056 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:47Z","lastTransitionTime":"2026-01-21T17:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:47 crc kubenswrapper[4808]: E0121 17:56:47.510696 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:47Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.516524 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.516573 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.516614 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.516631 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.516641 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:47Z","lastTransitionTime":"2026-01-21T17:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.519930 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.520008 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.520037 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.520217 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:56:47 crc kubenswrapper[4808]: E0121 17:56:47.520204 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:56:47 crc kubenswrapper[4808]: E0121 17:56:47.520413 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:56:47 crc kubenswrapper[4808]: E0121 17:56:47.520476 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:56:47 crc kubenswrapper[4808]: E0121 17:56:47.520528 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.522462 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 22:23:59.580794474 +0000 UTC Jan 21 17:56:47 crc kubenswrapper[4808]: E0121 17:56:47.539815 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:47Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:47 crc kubenswrapper[4808]: E0121 17:56:47.539934 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.541850 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.541917 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.541940 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.541973 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.541997 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:47Z","lastTransitionTime":"2026-01-21T17:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.646149 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.646197 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.646205 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.646222 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.646231 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:47Z","lastTransitionTime":"2026-01-21T17:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.749014 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.749067 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.749076 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.749097 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.749111 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:47Z","lastTransitionTime":"2026-01-21T17:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.851882 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.851933 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.851944 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.851962 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.851973 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:47Z","lastTransitionTime":"2026-01-21T17:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.955500 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.955574 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.955601 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.955633 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:47 crc kubenswrapper[4808]: I0121 17:56:47.955657 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:47Z","lastTransitionTime":"2026-01-21T17:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.058502 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.058575 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.058592 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.058614 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.058628 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:48Z","lastTransitionTime":"2026-01-21T17:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.161094 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.161197 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.161223 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.161288 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.161311 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:48Z","lastTransitionTime":"2026-01-21T17:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.270699 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.270772 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.270796 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.270826 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.270845 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:48Z","lastTransitionTime":"2026-01-21T17:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.373965 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.374058 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.374085 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.374118 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.374144 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:48Z","lastTransitionTime":"2026-01-21T17:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.476552 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.476595 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.476608 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.476626 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.476643 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:48Z","lastTransitionTime":"2026-01-21T17:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.523654 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 00:30:34.88317763 +0000 UTC Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.578912 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.578971 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.578986 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.579007 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.579021 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:48Z","lastTransitionTime":"2026-01-21T17:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.682122 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.682165 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.682176 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.682189 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.682198 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:48Z","lastTransitionTime":"2026-01-21T17:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.785795 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.785840 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.785850 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.785864 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.785875 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:48Z","lastTransitionTime":"2026-01-21T17:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.889036 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.889118 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.889195 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.889223 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.889268 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:48Z","lastTransitionTime":"2026-01-21T17:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.992935 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.992990 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.993011 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.993035 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:48 crc kubenswrapper[4808]: I0121 17:56:48.993050 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:48Z","lastTransitionTime":"2026-01-21T17:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.096439 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.096510 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.096533 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.096561 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.096584 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:49Z","lastTransitionTime":"2026-01-21T17:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.199896 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.199938 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.199948 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.199966 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.199978 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:49Z","lastTransitionTime":"2026-01-21T17:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.302923 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.302964 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.302978 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.303004 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.303018 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:49Z","lastTransitionTime":"2026-01-21T17:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.405892 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.405955 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.405975 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.406000 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.406017 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:49Z","lastTransitionTime":"2026-01-21T17:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.509497 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.509560 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.509579 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.509605 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.509622 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:49Z","lastTransitionTime":"2026-01-21T17:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.519764 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.519824 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.519787 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:56:49 crc kubenswrapper[4808]: E0121 17:56:49.519981 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:56:49 crc kubenswrapper[4808]: E0121 17:56:49.520105 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.520211 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:49 crc kubenswrapper[4808]: E0121 17:56:49.520361 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:56:49 crc kubenswrapper[4808]: E0121 17:56:49.520497 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.523855 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 15:33:02.886741206 +0000 UTC Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.613849 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.613911 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.613926 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.613948 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.613963 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:49Z","lastTransitionTime":"2026-01-21T17:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.717710 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.717800 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.717824 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.717857 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.717876 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:49Z","lastTransitionTime":"2026-01-21T17:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.820967 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.821015 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.821030 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.821050 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.821073 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:49Z","lastTransitionTime":"2026-01-21T17:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.923875 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.923927 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.923942 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.923962 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:49 crc kubenswrapper[4808]: I0121 17:56:49.923977 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:49Z","lastTransitionTime":"2026-01-21T17:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.027666 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.027713 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.027735 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.027757 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.027772 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:50Z","lastTransitionTime":"2026-01-21T17:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.130015 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.130076 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.130089 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.130107 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.130120 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:50Z","lastTransitionTime":"2026-01-21T17:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.232935 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.233004 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.233023 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.233054 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.233080 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:50Z","lastTransitionTime":"2026-01-21T17:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.336844 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.336910 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.336927 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.336952 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.336970 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:50Z","lastTransitionTime":"2026-01-21T17:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.440893 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.441087 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.441120 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.441151 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.441175 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:50Z","lastTransitionTime":"2026-01-21T17:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.524043 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 03:07:16.80398819 +0000 UTC Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.544598 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.544652 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.544664 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.544682 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.544694 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:50Z","lastTransitionTime":"2026-01-21T17:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.649138 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.649182 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.649195 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.649210 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.649225 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:50Z","lastTransitionTime":"2026-01-21T17:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.752367 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.752403 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.752413 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.752428 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.752438 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:50Z","lastTransitionTime":"2026-01-21T17:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.855181 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.855232 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.855274 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.855299 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.855318 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:50Z","lastTransitionTime":"2026-01-21T17:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.957834 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.957878 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.957890 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.957905 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:50 crc kubenswrapper[4808]: I0121 17:56:50.957918 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:50Z","lastTransitionTime":"2026-01-21T17:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.060004 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.060053 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.060069 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.060092 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.060108 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:51Z","lastTransitionTime":"2026-01-21T17:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.162285 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.162331 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.162355 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.162378 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.162394 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:51Z","lastTransitionTime":"2026-01-21T17:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.195355 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.204137 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.207663 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:51Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.223304 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:51Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.235356 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:51Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.246487 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:51Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.255895 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs\") pod \"network-metrics-daemon-gm2t2\" (UID: \"8d39c565-f404-451e-896f-ecc5af76a000\") " pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:56:51 crc kubenswrapper[4808]: E0121 17:56:51.256017 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 17:56:51 crc kubenswrapper[4808]: E0121 17:56:51.256080 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs podName:8d39c565-f404-451e-896f-ecc5af76a000 nodeName:}" failed. No retries permitted until 2026-01-21 17:56:59.256064457 +0000 UTC m=+53.986575342 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs") pod "network-metrics-daemon-gm2t2" (UID: "8d39c565-f404-451e-896f-ecc5af76a000") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.258075 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53885bfd-41c1-49a4-beb3-e937534a6ef8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce09f9763152c19c6f0cc141317c798d49f50066d1f0184a114085a19fc9adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90dfe0143cdd3cd582497eddee45eeb42dd96a03eb26a3b753c8aa2fb37436bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw46g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:51Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.265517 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.265545 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.265558 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.265576 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.265589 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:51Z","lastTransitionTime":"2026-01-21T17:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.275464 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:51Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.291241 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:51Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.307927 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:51Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.322487 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:51Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.345373 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:51Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.360012 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:51Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.368365 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.368399 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.368410 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.368426 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.368438 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:51Z","lastTransitionTime":"2026-01-21T17:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.370236 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:51Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.390729 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:51Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.407592 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:51Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.426422 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:40Z\\\",\\\"message\\\":\\\"ectMeta:{metrics openshift-etcd-operator ff1da138-ae82-4792-ae1f-3b2df1427723 4289 0 2025-02-23 05:12:19 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:etcd-operator] map[include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-secret-name:etcd-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00078f0a7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: etcd-operator,},ClusterIP:10.217.5.188,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.188],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Conditi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fc2fk_openshift-ovn-kubernetes(bb13a20f-f4ab-412c-8165-5b93edc79628)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:51Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.438847 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gm2t2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d39c565-f404-451e-896f-ecc5af76a000\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gm2t2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:51Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.470525 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.470565 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.470574 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.470588 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.470597 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:51Z","lastTransitionTime":"2026-01-21T17:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.521528 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.521658 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.521774 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.521812 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:56:51 crc kubenswrapper[4808]: E0121 17:56:51.521699 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:56:51 crc kubenswrapper[4808]: E0121 17:56:51.521852 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:56:51 crc kubenswrapper[4808]: E0121 17:56:51.521903 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:56:51 crc kubenswrapper[4808]: E0121 17:56:51.522020 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.524651 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 02:04:33.594823822 +0000 UTC Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.572362 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.572403 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.572415 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.572433 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.572445 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:51Z","lastTransitionTime":"2026-01-21T17:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.675885 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.675924 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.675935 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.675950 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.675962 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:51Z","lastTransitionTime":"2026-01-21T17:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.778758 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.778825 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.778853 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.778881 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.778903 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:51Z","lastTransitionTime":"2026-01-21T17:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.881971 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.882062 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.882091 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.882120 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.882139 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:51Z","lastTransitionTime":"2026-01-21T17:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.984362 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.984394 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.984403 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.984416 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:51 crc kubenswrapper[4808]: I0121 17:56:51.984425 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:51Z","lastTransitionTime":"2026-01-21T17:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.086443 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.086489 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.086503 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.086525 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.086540 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:52Z","lastTransitionTime":"2026-01-21T17:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.189981 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.190280 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.190374 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.190472 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.190588 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:52Z","lastTransitionTime":"2026-01-21T17:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.294963 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.295002 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.295014 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.295029 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.295039 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:52Z","lastTransitionTime":"2026-01-21T17:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.398042 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.398100 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.398115 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.398136 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.398151 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:52Z","lastTransitionTime":"2026-01-21T17:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.500809 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.500866 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.500883 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.500898 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.500908 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:52Z","lastTransitionTime":"2026-01-21T17:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.525165 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 21:27:55.852585086 +0000 UTC Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.603323 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.603363 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.603373 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.603389 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.603406 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:52Z","lastTransitionTime":"2026-01-21T17:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.706009 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.706070 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.706086 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.706107 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.706122 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:52Z","lastTransitionTime":"2026-01-21T17:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.808281 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.808325 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.808338 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.808352 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.808361 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:52Z","lastTransitionTime":"2026-01-21T17:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.911091 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.911141 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.911157 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.911178 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:52 crc kubenswrapper[4808]: I0121 17:56:52.911193 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:52Z","lastTransitionTime":"2026-01-21T17:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.014458 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.014523 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.014540 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.014568 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.014585 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:53Z","lastTransitionTime":"2026-01-21T17:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.117218 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.117285 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.117294 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.117312 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.117324 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:53Z","lastTransitionTime":"2026-01-21T17:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.220751 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.220824 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.220862 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.220900 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.220923 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:53Z","lastTransitionTime":"2026-01-21T17:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.323308 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.323379 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.323400 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.323422 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.323431 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:53Z","lastTransitionTime":"2026-01-21T17:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.426779 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.426857 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.426880 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.426908 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.426929 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:53Z","lastTransitionTime":"2026-01-21T17:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.520392 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:53 crc kubenswrapper[4808]: E0121 17:56:53.520579 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.520686 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.520707 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.520841 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:53 crc kubenswrapper[4808]: E0121 17:56:53.520993 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:56:53 crc kubenswrapper[4808]: E0121 17:56:53.521136 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:56:53 crc kubenswrapper[4808]: E0121 17:56:53.521318 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.526809 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 15:43:38.265691651 +0000 UTC Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.531723 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.531820 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.531848 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.531881 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.531905 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:53Z","lastTransitionTime":"2026-01-21T17:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.635825 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.635888 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.635906 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.635930 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.635946 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:53Z","lastTransitionTime":"2026-01-21T17:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.738723 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.738836 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.738857 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.738886 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.738906 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:53Z","lastTransitionTime":"2026-01-21T17:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.842449 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.842603 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.842632 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.842715 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.842786 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:53Z","lastTransitionTime":"2026-01-21T17:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.945789 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.945869 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.945892 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.945920 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:53 crc kubenswrapper[4808]: I0121 17:56:53.945938 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:53Z","lastTransitionTime":"2026-01-21T17:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.049782 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.049850 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.049888 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.049923 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.049950 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:54Z","lastTransitionTime":"2026-01-21T17:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.152665 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.152745 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.152761 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.152784 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.152802 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:54Z","lastTransitionTime":"2026-01-21T17:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.255955 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.256025 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.256039 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.256062 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.256079 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:54Z","lastTransitionTime":"2026-01-21T17:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.359764 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.359832 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.359851 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.359878 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.359897 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:54Z","lastTransitionTime":"2026-01-21T17:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.463744 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.463821 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.463838 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.463865 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.463884 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:54Z","lastTransitionTime":"2026-01-21T17:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.527291 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 09:13:46.201299884 +0000 UTC Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.567590 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.567656 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.567673 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.567698 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.567717 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:54Z","lastTransitionTime":"2026-01-21T17:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.671139 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.671218 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.671279 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.671315 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.671347 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:54Z","lastTransitionTime":"2026-01-21T17:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.774951 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.775005 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.775022 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.775046 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.775062 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:54Z","lastTransitionTime":"2026-01-21T17:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.878496 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.878579 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.878615 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.878653 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.878680 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:54Z","lastTransitionTime":"2026-01-21T17:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.981102 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.981172 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.981199 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.981230 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:54 crc kubenswrapper[4808]: I0121 17:56:54.981294 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:54Z","lastTransitionTime":"2026-01-21T17:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.084423 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.084498 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.084523 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.084556 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.084579 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:55Z","lastTransitionTime":"2026-01-21T17:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.187600 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.187674 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.187696 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.187723 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.187743 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:55Z","lastTransitionTime":"2026-01-21T17:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.290768 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.291038 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.291102 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.291267 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.291350 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:55Z","lastTransitionTime":"2026-01-21T17:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.394055 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.394104 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.394116 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.394130 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.394141 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:55Z","lastTransitionTime":"2026-01-21T17:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.404948 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:56:55 crc kubenswrapper[4808]: E0121 17:56:55.405106 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:57:27.405088311 +0000 UTC m=+82.135599196 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.497491 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.497545 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.497556 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.497583 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.497595 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:55Z","lastTransitionTime":"2026-01-21T17:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.506131 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.506204 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.506301 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.506362 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:55 crc kubenswrapper[4808]: E0121 17:56:55.506545 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 17:56:55 crc kubenswrapper[4808]: E0121 17:56:55.506580 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 17:56:55 crc kubenswrapper[4808]: E0121 17:56:55.506629 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 17:57:27.506603084 +0000 UTC m=+82.237114009 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 17:56:55 crc kubenswrapper[4808]: E0121 17:56:55.506673 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 17:57:27.506641985 +0000 UTC m=+82.237152960 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 17:56:55 crc kubenswrapper[4808]: E0121 17:56:55.506699 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 17:56:55 crc kubenswrapper[4808]: E0121 17:56:55.506731 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 17:56:55 crc kubenswrapper[4808]: E0121 17:56:55.506745 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:55 crc kubenswrapper[4808]: E0121 17:56:55.506803 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 17:57:27.506785858 +0000 UTC m=+82.237296743 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:55 crc kubenswrapper[4808]: E0121 17:56:55.506915 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 17:56:55 crc kubenswrapper[4808]: E0121 17:56:55.506993 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 17:56:55 crc kubenswrapper[4808]: E0121 17:56:55.507079 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:55 crc kubenswrapper[4808]: E0121 17:56:55.507175 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 17:57:27.507165556 +0000 UTC m=+82.237676431 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.519922 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.519991 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:55 crc kubenswrapper[4808]: E0121 17:56:55.520312 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.520061 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:55 crc kubenswrapper[4808]: E0121 17:56:55.520694 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:56:55 crc kubenswrapper[4808]: E0121 17:56:55.520495 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.520059 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:55 crc kubenswrapper[4808]: E0121 17:56:55.521061 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.522110 4808 scope.go:117] "RemoveContainer" containerID="019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.527507 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 07:33:42.774045955 +0000 UTC Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.534562 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.569806 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:40Z\\\",\\\"message\\\":\\\"ectMeta:{metrics openshift-etcd-operator ff1da138-ae82-4792-ae1f-3b2df1427723 4289 0 2025-02-23 05:12:19 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:etcd-operator] map[include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-secret-name:etcd-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00078f0a7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: etcd-operator,},ClusterIP:10.217.5.188,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.188],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Conditi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fc2fk_openshift-ovn-kubernetes(bb13a20f-f4ab-412c-8165-5b93edc79628)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.585718 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gm2t2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d39c565-f404-451e-896f-ecc5af76a000\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gm2t2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.599078 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.599112 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.599122 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.599136 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.599145 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:55Z","lastTransitionTime":"2026-01-21T17:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.603630 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.618517 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b148e21a-1611-43f8-85c3-ac3899dde92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65ea8da4528954ea712538a48e6e1c6a1fe893873742ad120f6300ee32f6c95b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59727192447c6eef4b045ac5a9a0c0a7bba9c0f2672d46ff1fc9b2b58c022279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94585d6e08b85c96da3d90042ca20e0a194e7488185ef798b4c2dcc2c2948897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.631997 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.647218 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.662920 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.676098 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53885bfd-41c1-49a4-beb3-e937534a6ef8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce09f9763152c19c6f0cc141317c798d49f50066d1f0184a114085a19fc9adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90dfe0143cdd3cd582497eddee45eeb42dd96a03eb26a3b753c8aa2fb37436bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw46g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.689678 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.702920 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.702982 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.703000 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.703027 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.703047 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:55Z","lastTransitionTime":"2026-01-21T17:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.703376 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.719770 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.737104 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.757124 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.778599 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.791905 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.806465 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.806514 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.806526 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.806546 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.806558 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:55Z","lastTransitionTime":"2026-01-21T17:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.806949 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.831418 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovnkube-controller/1.log" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.835433 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerStarted","Data":"b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f"} Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.836696 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.851926 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.864661 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53885bfd-41c1-49a4-beb3-e937534a6ef8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce09f9763152c19c6f0cc141317c798d49f50066d1f0184a114085a19fc9adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90dfe0143cdd3cd582497eddee45eeb42dd96a03eb26a3b753c8aa2fb37436bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw46g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.882303 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.902801 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.910419 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.910444 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.910452 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.910467 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.910478 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:55Z","lastTransitionTime":"2026-01-21T17:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.926785 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.946001 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.966525 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:55 crc kubenswrapper[4808]: I0121 17:56:55.988143 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.002112 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:55Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.014655 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.014692 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.014704 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.014724 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.014739 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:56Z","lastTransitionTime":"2026-01-21T17:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.021290 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b148e21a-1611-43f8-85c3-ac3899dde92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65ea8da4528954ea712538a48e6e1c6a1fe893873742ad120f6300ee32f6c95b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59727192447c6eef4b045ac5a9a0c0a7bba9c0f2672d46ff1fc9b2b58c022279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94585d6e08b85c96da3d90042ca20e0a194e7488185ef798b4c2dcc2c2948897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:56Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.037167 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:56Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.056809 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:40Z\\\",\\\"message\\\":\\\"ectMeta:{metrics openshift-etcd-operator ff1da138-ae82-4792-ae1f-3b2df1427723 4289 0 2025-02-23 05:12:19 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:etcd-operator] map[include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-secret-name:etcd-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00078f0a7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: etcd-operator,},ClusterIP:10.217.5.188,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.188],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Conditi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:56Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.067554 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gm2t2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d39c565-f404-451e-896f-ecc5af76a000\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gm2t2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:56Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.082006 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:56Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.095056 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:56Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.108861 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:56Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.117746 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.117804 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.117818 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.117842 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.117854 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:56Z","lastTransitionTime":"2026-01-21T17:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.122193 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:56Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.220093 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.220145 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.220160 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.220186 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.220202 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:56Z","lastTransitionTime":"2026-01-21T17:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.322450 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.322482 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.322494 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.322508 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.322518 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:56Z","lastTransitionTime":"2026-01-21T17:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.424726 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.424761 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.424773 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.424790 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.424803 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:56Z","lastTransitionTime":"2026-01-21T17:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.527946 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 11:09:33.543840782 +0000 UTC Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.528076 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.528110 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.528126 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.528153 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.528171 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:56Z","lastTransitionTime":"2026-01-21T17:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.632084 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.632163 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.632191 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.632222 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.632275 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:56Z","lastTransitionTime":"2026-01-21T17:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.735789 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.735857 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.735880 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.735906 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.735921 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:56Z","lastTransitionTime":"2026-01-21T17:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.838545 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.838599 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.838609 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.838626 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.838638 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:56Z","lastTransitionTime":"2026-01-21T17:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.941574 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.941982 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.942110 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.942231 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:56 crc kubenswrapper[4808]: I0121 17:56:56.942398 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:56Z","lastTransitionTime":"2026-01-21T17:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.045172 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.045528 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.045632 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.045719 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.045795 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:57Z","lastTransitionTime":"2026-01-21T17:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.148832 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.149064 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.149160 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.149223 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.149326 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:57Z","lastTransitionTime":"2026-01-21T17:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.253614 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.253937 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.254043 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.254159 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.254287 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:57Z","lastTransitionTime":"2026-01-21T17:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.357423 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.357854 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.357933 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.357998 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.358068 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:57Z","lastTransitionTime":"2026-01-21T17:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.461672 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.461739 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.461751 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.461790 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.461807 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:57Z","lastTransitionTime":"2026-01-21T17:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.519973 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.520082 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:57 crc kubenswrapper[4808]: E0121 17:56:57.520129 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.520183 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.520195 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:56:57 crc kubenswrapper[4808]: E0121 17:56:57.520366 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:56:57 crc kubenswrapper[4808]: E0121 17:56:57.520499 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:56:57 crc kubenswrapper[4808]: E0121 17:56:57.520589 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.528424 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 14:50:34.62421198 +0000 UTC Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.564384 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.564435 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.564453 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.564474 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.564492 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:57Z","lastTransitionTime":"2026-01-21T17:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.667408 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.667475 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.667515 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.667542 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.667558 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:57Z","lastTransitionTime":"2026-01-21T17:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.708897 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.708949 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.708959 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.708976 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.708987 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:57Z","lastTransitionTime":"2026-01-21T17:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:57 crc kubenswrapper[4808]: E0121 17:56:57.724044 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:57Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.727976 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.728035 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.728046 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.728072 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.728085 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:57Z","lastTransitionTime":"2026-01-21T17:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:57 crc kubenswrapper[4808]: E0121 17:56:57.749325 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:57Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.754711 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.754763 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.754775 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.754797 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.754807 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:57Z","lastTransitionTime":"2026-01-21T17:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:57 crc kubenswrapper[4808]: E0121 17:56:57.767804 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:57Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.772930 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.773006 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.773020 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.773044 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.773062 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:57Z","lastTransitionTime":"2026-01-21T17:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:57 crc kubenswrapper[4808]: E0121 17:56:57.790599 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:57Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.795371 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.795515 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.795534 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.795556 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.795571 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:57Z","lastTransitionTime":"2026-01-21T17:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:57 crc kubenswrapper[4808]: E0121 17:56:57.809458 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:57Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:57 crc kubenswrapper[4808]: E0121 17:56:57.809622 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.811053 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.811090 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.811103 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.811120 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.811133 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:57Z","lastTransitionTime":"2026-01-21T17:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.843980 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovnkube-controller/2.log" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.844650 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovnkube-controller/1.log" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.848430 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerDied","Data":"b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f"} Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.848431 4808 generic.go:334] "Generic (PLEG): container finished" podID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerID="b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f" exitCode=1 Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.848528 4808 scope.go:117] "RemoveContainer" containerID="019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.849162 4808 scope.go:117] "RemoveContainer" containerID="b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f" Jan 21 17:56:57 crc kubenswrapper[4808]: E0121 17:56:57.849319 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fc2fk_openshift-ovn-kubernetes(bb13a20f-f4ab-412c-8165-5b93edc79628)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.864770 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:57Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.878703 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:57Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.893156 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:57Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.908184 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:57Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.913620 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.913677 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.913696 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.913724 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.913742 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:57Z","lastTransitionTime":"2026-01-21T17:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.920289 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:57Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.934696 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:57Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.953356 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:57Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.971596 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:57Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:57 crc kubenswrapper[4808]: I0121 17:56:57.984414 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b148e21a-1611-43f8-85c3-ac3899dde92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65ea8da4528954ea712538a48e6e1c6a1fe893873742ad120f6300ee32f6c95b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59727192447c6eef4b045ac5a9a0c0a7bba9c0f2672d46ff1fc9b2b58c022279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94585d6e08b85c96da3d90042ca20e0a194e7488185ef798b4c2dcc2c2948897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:57Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.000581 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:57Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.016832 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.016879 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.016888 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.016907 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.016918 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:58Z","lastTransitionTime":"2026-01-21T17:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.024879 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:40Z\\\",\\\"message\\\":\\\"ectMeta:{metrics openshift-etcd-operator ff1da138-ae82-4792-ae1f-3b2df1427723 4289 0 2025-02-23 05:12:19 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:etcd-operator] map[include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-secret-name:etcd-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00078f0a7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: etcd-operator,},ClusterIP:10.217.5.188,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.188],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Conditi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"ble:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:static_routes Mutator:insert Value:{GoSet:[{GoUUID:8944024f-deb7-4076-afb3-4b50a2ff4b4b}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 17:56:56.395979 6462 obj_retry.go:551] Creating *factory.egressNode crc took: 2.120476ms\\\\nI0121 17:56:56.396002 6462 factory.go:1336] Added *v1.Node event handler 7\\\\nI0121 17:56:56.396015 6462 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0121 17:56:56.396027 6462 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0121 17:56:56.396029 6462 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0121 17:56:56.396062 6462 factory.go:656] Stopping watch factory\\\\nI0121 17:56:56.396086 6462 handler.go:208] Removed *v1.Node event handler 7\\\\nI0121 17:56:56.396134 6462 handler.go:208] Removed *v1.Node event handler 2\\\\nI0121 17:56:56.396285 6462 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0121 17:56:56.396360 6462 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0121 17:56:56.396393 6462 ovnkube.go:599] Stopped ovnkube\\\\nI0121 17:56:56.396421 6462 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0121 17:56:56.396480 6462 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:58Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.039634 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gm2t2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d39c565-f404-451e-896f-ecc5af76a000\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gm2t2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:58Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.058230 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:58Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.075340 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:58Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.089190 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:58Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.103708 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:58Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.118969 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53885bfd-41c1-49a4-beb3-e937534a6ef8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce09f9763152c19c6f0cc141317c798d49f50066d1f0184a114085a19fc9adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90dfe0143cdd3cd582497eddee45eeb42dd96a03eb26a3b753c8aa2fb37436bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw46g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:56:58Z is after 2025-08-24T17:21:41Z" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.119535 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.119572 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.119587 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.119610 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.119625 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:58Z","lastTransitionTime":"2026-01-21T17:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.221761 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.221833 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.221851 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.221875 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.221893 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:58Z","lastTransitionTime":"2026-01-21T17:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.324967 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.325495 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.325648 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.325791 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.325936 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:58Z","lastTransitionTime":"2026-01-21T17:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.428757 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.428819 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.428837 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.428864 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.428889 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:58Z","lastTransitionTime":"2026-01-21T17:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.529086 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 17:48:01.207687671 +0000 UTC Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.531156 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.531208 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.531225 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.531283 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.531301 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:58Z","lastTransitionTime":"2026-01-21T17:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.635783 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.635814 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.635823 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.635836 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.635844 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:58Z","lastTransitionTime":"2026-01-21T17:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.741870 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.741918 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.741927 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.741942 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.741951 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:58Z","lastTransitionTime":"2026-01-21T17:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.844731 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.844775 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.844786 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.844801 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.844811 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:58Z","lastTransitionTime":"2026-01-21T17:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.852584 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovnkube-controller/2.log" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.947428 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.947474 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.947485 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.947501 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:58 crc kubenswrapper[4808]: I0121 17:56:58.947512 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:58Z","lastTransitionTime":"2026-01-21T17:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.049471 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.049515 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.049525 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.049540 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.049552 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:59Z","lastTransitionTime":"2026-01-21T17:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.151961 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.152000 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.152011 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.152027 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.152036 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:59Z","lastTransitionTime":"2026-01-21T17:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.254061 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.254424 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.254434 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.254449 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.254457 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:59Z","lastTransitionTime":"2026-01-21T17:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.276956 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs\") pod \"network-metrics-daemon-gm2t2\" (UID: \"8d39c565-f404-451e-896f-ecc5af76a000\") " pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:56:59 crc kubenswrapper[4808]: E0121 17:56:59.277176 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 17:56:59 crc kubenswrapper[4808]: E0121 17:56:59.277338 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs podName:8d39c565-f404-451e-896f-ecc5af76a000 nodeName:}" failed. No retries permitted until 2026-01-21 17:57:15.277306396 +0000 UTC m=+70.007817311 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs") pod "network-metrics-daemon-gm2t2" (UID: "8d39c565-f404-451e-896f-ecc5af76a000") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.356796 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.356830 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.356842 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.356858 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.356868 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:59Z","lastTransitionTime":"2026-01-21T17:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.459787 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.459842 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.459857 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.459877 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.459893 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:59Z","lastTransitionTime":"2026-01-21T17:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.520566 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:56:59 crc kubenswrapper[4808]: E0121 17:56:59.520691 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.520807 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.520578 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:56:59 crc kubenswrapper[4808]: E0121 17:56:59.520922 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:56:59 crc kubenswrapper[4808]: E0121 17:56:59.520996 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.521411 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:56:59 crc kubenswrapper[4808]: E0121 17:56:59.521977 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.529531 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 13:57:59.215500553 +0000 UTC Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.563596 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.563652 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.563670 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.563725 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.563764 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:59Z","lastTransitionTime":"2026-01-21T17:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.665899 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.665940 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.665950 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.665968 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.665980 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:59Z","lastTransitionTime":"2026-01-21T17:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.769177 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.769221 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.769229 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.769272 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.769283 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:59Z","lastTransitionTime":"2026-01-21T17:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.871201 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.871269 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.871283 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.871298 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.871310 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:59Z","lastTransitionTime":"2026-01-21T17:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.974144 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.974595 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.974790 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.974988 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:56:59 crc kubenswrapper[4808]: I0121 17:56:59.975119 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:56:59Z","lastTransitionTime":"2026-01-21T17:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.078569 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.078660 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.078677 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.078702 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.078719 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:00Z","lastTransitionTime":"2026-01-21T17:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.182115 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.182183 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.182229 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.182306 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.182395 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:00Z","lastTransitionTime":"2026-01-21T17:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.286073 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.286111 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.286121 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.286136 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.286145 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:00Z","lastTransitionTime":"2026-01-21T17:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.389009 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.389366 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.389457 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.389554 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.389647 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:00Z","lastTransitionTime":"2026-01-21T17:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.492236 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.492591 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.492684 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.492784 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.492882 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:00Z","lastTransitionTime":"2026-01-21T17:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.529878 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 21:47:44.513757264 +0000 UTC Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.595202 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.595364 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.595424 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.595503 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.595577 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:00Z","lastTransitionTime":"2026-01-21T17:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.698563 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.698619 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.698634 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.698656 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.698671 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:00Z","lastTransitionTime":"2026-01-21T17:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.801309 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.801369 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.801391 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.801421 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.801443 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:00Z","lastTransitionTime":"2026-01-21T17:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.903738 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.903800 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.903819 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.903847 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:00 crc kubenswrapper[4808]: I0121 17:57:00.903871 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:00Z","lastTransitionTime":"2026-01-21T17:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.009110 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.009165 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.009183 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.009206 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.009224 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:01Z","lastTransitionTime":"2026-01-21T17:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.111686 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.111746 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.111763 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.111787 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.111804 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:01Z","lastTransitionTime":"2026-01-21T17:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.214728 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.214771 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.214786 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.214811 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.214827 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:01Z","lastTransitionTime":"2026-01-21T17:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.317614 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.317677 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.317693 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.317718 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.317736 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:01Z","lastTransitionTime":"2026-01-21T17:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.420874 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.420947 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.420971 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.421000 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.421021 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:01Z","lastTransitionTime":"2026-01-21T17:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.520484 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.520574 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.520632 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:01 crc kubenswrapper[4808]: E0121 17:57:01.520706 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.520752 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:01 crc kubenswrapper[4808]: E0121 17:57:01.520908 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:01 crc kubenswrapper[4808]: E0121 17:57:01.521089 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:01 crc kubenswrapper[4808]: E0121 17:57:01.521303 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.523108 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.523149 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.523159 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.523175 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.523187 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:01Z","lastTransitionTime":"2026-01-21T17:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.530782 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 23:13:40.779002691 +0000 UTC Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.625694 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.625754 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.625771 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.625797 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.625814 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:01Z","lastTransitionTime":"2026-01-21T17:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.728173 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.728238 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.728281 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.728308 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.728327 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:01Z","lastTransitionTime":"2026-01-21T17:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.831335 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.831399 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.831417 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.831443 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.831461 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:01Z","lastTransitionTime":"2026-01-21T17:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.935071 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.935179 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.935205 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.935311 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:01 crc kubenswrapper[4808]: I0121 17:57:01.935343 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:01Z","lastTransitionTime":"2026-01-21T17:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.038166 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.038273 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.038298 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.038328 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.038351 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:02Z","lastTransitionTime":"2026-01-21T17:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.141765 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.141881 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.141908 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.141939 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.141960 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:02Z","lastTransitionTime":"2026-01-21T17:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.244886 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.244957 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.244971 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.244995 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.245010 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:02Z","lastTransitionTime":"2026-01-21T17:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.348997 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.349061 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.349078 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.349103 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.349118 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:02Z","lastTransitionTime":"2026-01-21T17:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.452950 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.453005 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.453026 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.453051 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.453066 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:02Z","lastTransitionTime":"2026-01-21T17:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.530934 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 20:28:43.723445864 +0000 UTC Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.557079 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.557150 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.557172 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.557201 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.557227 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:02Z","lastTransitionTime":"2026-01-21T17:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.660117 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.660171 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.660188 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.660214 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.660274 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:02Z","lastTransitionTime":"2026-01-21T17:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.764559 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.764642 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.764661 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.764689 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.764706 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:02Z","lastTransitionTime":"2026-01-21T17:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.868140 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.868202 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.868226 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.868296 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.868323 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:02Z","lastTransitionTime":"2026-01-21T17:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.972057 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.972136 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.972152 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.972179 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:02 crc kubenswrapper[4808]: I0121 17:57:02.972198 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:02Z","lastTransitionTime":"2026-01-21T17:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.076325 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.076394 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.076418 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.076450 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.076472 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:03Z","lastTransitionTime":"2026-01-21T17:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.181902 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.182033 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.182062 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.182097 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.182123 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:03Z","lastTransitionTime":"2026-01-21T17:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.285939 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.285998 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.286019 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.286046 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.286064 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:03Z","lastTransitionTime":"2026-01-21T17:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.389740 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.389845 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.389865 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.389895 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.389912 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:03Z","lastTransitionTime":"2026-01-21T17:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.493888 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.493959 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.493977 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.494003 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.494022 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:03Z","lastTransitionTime":"2026-01-21T17:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.520515 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:03 crc kubenswrapper[4808]: E0121 17:57:03.520752 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.521472 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:03 crc kubenswrapper[4808]: E0121 17:57:03.521601 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.521658 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.521666 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:03 crc kubenswrapper[4808]: E0121 17:57:03.521783 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:03 crc kubenswrapper[4808]: E0121 17:57:03.524501 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.532093 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 18:29:20.809943603 +0000 UTC Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.596526 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.596602 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.596623 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.596656 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.596678 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:03Z","lastTransitionTime":"2026-01-21T17:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.699681 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.699737 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.699755 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.699782 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.699801 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:03Z","lastTransitionTime":"2026-01-21T17:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.803141 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.803214 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.803232 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.803318 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.803339 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:03Z","lastTransitionTime":"2026-01-21T17:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.906503 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.906589 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.906614 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.906647 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:03 crc kubenswrapper[4808]: I0121 17:57:03.906674 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:03Z","lastTransitionTime":"2026-01-21T17:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.009618 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.009696 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.009714 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.009740 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.009757 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:04Z","lastTransitionTime":"2026-01-21T17:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.113368 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.113440 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.113463 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.113535 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.113560 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:04Z","lastTransitionTime":"2026-01-21T17:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.216820 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.216872 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.216889 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.216911 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.216927 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:04Z","lastTransitionTime":"2026-01-21T17:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.320221 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.320327 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.320368 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.320402 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.320421 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:04Z","lastTransitionTime":"2026-01-21T17:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.423824 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.423877 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.423893 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.423914 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.423930 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:04Z","lastTransitionTime":"2026-01-21T17:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.527344 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.527413 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.527431 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.527454 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.527495 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:04Z","lastTransitionTime":"2026-01-21T17:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.532811 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 10:07:16.402423628 +0000 UTC Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.629558 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.629584 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.629592 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.629607 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.629624 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:04Z","lastTransitionTime":"2026-01-21T17:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.731971 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.732024 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.732037 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.732059 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.732074 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:04Z","lastTransitionTime":"2026-01-21T17:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.835442 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.835518 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.835560 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.835599 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.835626 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:04Z","lastTransitionTime":"2026-01-21T17:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.939101 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.939188 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.939221 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.939291 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:04 crc kubenswrapper[4808]: I0121 17:57:04.939317 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:04Z","lastTransitionTime":"2026-01-21T17:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.042768 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.042830 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.042857 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.042886 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.042909 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:05Z","lastTransitionTime":"2026-01-21T17:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.146529 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.146571 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.146581 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.146597 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.146609 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:05Z","lastTransitionTime":"2026-01-21T17:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.249708 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.249765 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.249775 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.249792 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.249801 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:05Z","lastTransitionTime":"2026-01-21T17:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.352392 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.352453 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.352478 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.352510 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.352532 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:05Z","lastTransitionTime":"2026-01-21T17:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.455240 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.455356 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.455375 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.455401 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.455418 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:05Z","lastTransitionTime":"2026-01-21T17:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.520121 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.520228 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.520121 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:05 crc kubenswrapper[4808]: E0121 17:57:05.520427 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:05 crc kubenswrapper[4808]: E0121 17:57:05.520605 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:05 crc kubenswrapper[4808]: E0121 17:57:05.520786 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.521050 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:05 crc kubenswrapper[4808]: E0121 17:57:05.521152 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.533107 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 18:18:14.445908194 +0000 UTC Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.536783 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:05Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.557032 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:05Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.561155 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.561198 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.561207 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.561223 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.561233 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:05Z","lastTransitionTime":"2026-01-21T17:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.574560 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:05Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.590972 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:05Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.608203 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:05Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.625963 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:05Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.642378 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:05Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.659786 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:05Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.663932 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.663978 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.663991 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.664013 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.664028 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:05Z","lastTransitionTime":"2026-01-21T17:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.673002 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b148e21a-1611-43f8-85c3-ac3899dde92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65ea8da4528954ea712538a48e6e1c6a1fe893873742ad120f6300ee32f6c95b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59727192447c6eef4b045ac5a9a0c0a7bba9c0f2672d46ff1fc9b2b58c022279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94585d6e08b85c96da3d90042ca20e0a194e7488185ef798b4c2dcc2c2948897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:05Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.686571 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:05Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.708171 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019d9279a130a598d27507afd5a70a91b7c05be8fda8687c6b06d16866d8e3a0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:40Z\\\",\\\"message\\\":\\\"ectMeta:{metrics openshift-etcd-operator ff1da138-ae82-4792-ae1f-3b2df1427723 4289 0 2025-02-23 05:12:19 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:etcd-operator] map[include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-secret-name:etcd-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00078f0a7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: etcd-operator,},ClusterIP:10.217.5.188,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.188],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Conditi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"ble:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:static_routes Mutator:insert Value:{GoSet:[{GoUUID:8944024f-deb7-4076-afb3-4b50a2ff4b4b}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 17:56:56.395979 6462 obj_retry.go:551] Creating *factory.egressNode crc took: 2.120476ms\\\\nI0121 17:56:56.396002 6462 factory.go:1336] Added *v1.Node event handler 7\\\\nI0121 17:56:56.396015 6462 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0121 17:56:56.396027 6462 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0121 17:56:56.396029 6462 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0121 17:56:56.396062 6462 factory.go:656] Stopping watch factory\\\\nI0121 17:56:56.396086 6462 handler.go:208] Removed *v1.Node event handler 7\\\\nI0121 17:56:56.396134 6462 handler.go:208] Removed *v1.Node event handler 2\\\\nI0121 17:56:56.396285 6462 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0121 17:56:56.396360 6462 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0121 17:56:56.396393 6462 ovnkube.go:599] Stopped ovnkube\\\\nI0121 17:56:56.396421 6462 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0121 17:56:56.396480 6462 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:05Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.722816 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gm2t2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d39c565-f404-451e-896f-ecc5af76a000\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gm2t2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:05Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.737744 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:05Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.754648 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:05Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.766209 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.766296 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.766309 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.766329 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.766343 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:05Z","lastTransitionTime":"2026-01-21T17:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.768792 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:05Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.783678 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:05Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.798995 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53885bfd-41c1-49a4-beb3-e937534a6ef8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce09f9763152c19c6f0cc141317c798d49f50066d1f0184a114085a19fc9adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90dfe0143cdd3cd582497eddee45eeb42dd96a03eb26a3b753c8aa2fb37436bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw46g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:05Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.874634 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.874706 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.874735 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.874766 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.874797 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:05Z","lastTransitionTime":"2026-01-21T17:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.978858 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.979004 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.979056 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.979083 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:05 crc kubenswrapper[4808]: I0121 17:57:05.979129 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:05Z","lastTransitionTime":"2026-01-21T17:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.082513 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.082989 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.083006 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.083022 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.083033 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:06Z","lastTransitionTime":"2026-01-21T17:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.186126 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.186212 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.186567 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.186616 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.186641 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:06Z","lastTransitionTime":"2026-01-21T17:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.290082 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.290139 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.290156 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.290179 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.290195 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:06Z","lastTransitionTime":"2026-01-21T17:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.394137 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.394198 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.394303 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.394375 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.394405 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:06Z","lastTransitionTime":"2026-01-21T17:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.497602 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.497673 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.497687 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.497714 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.497732 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:06Z","lastTransitionTime":"2026-01-21T17:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.533487 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 08:00:44.808811623 +0000 UTC Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.601037 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.601119 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.601141 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.601166 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.601181 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:06Z","lastTransitionTime":"2026-01-21T17:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.703964 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.704049 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.704077 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.704108 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.704131 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:06Z","lastTransitionTime":"2026-01-21T17:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.807343 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.807391 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.807417 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.807435 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.807445 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:06Z","lastTransitionTime":"2026-01-21T17:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.910348 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.910425 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.910451 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.910483 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:06 crc kubenswrapper[4808]: I0121 17:57:06.910507 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:06Z","lastTransitionTime":"2026-01-21T17:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.013825 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.013876 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.013894 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.013919 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.013936 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:07Z","lastTransitionTime":"2026-01-21T17:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.116828 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.116877 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.116890 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.116908 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.116922 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:07Z","lastTransitionTime":"2026-01-21T17:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.219860 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.220079 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.220086 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.220101 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.220110 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:07Z","lastTransitionTime":"2026-01-21T17:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.322129 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.322191 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.322206 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.322225 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.322237 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:07Z","lastTransitionTime":"2026-01-21T17:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.425301 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.425347 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.425362 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.425381 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.425391 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:07Z","lastTransitionTime":"2026-01-21T17:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.520192 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.520277 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.520295 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.520212 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:07 crc kubenswrapper[4808]: E0121 17:57:07.520401 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:07 crc kubenswrapper[4808]: E0121 17:57:07.520525 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:07 crc kubenswrapper[4808]: E0121 17:57:07.520627 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:07 crc kubenswrapper[4808]: E0121 17:57:07.520704 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.527293 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.527344 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.527360 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.527381 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.527396 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:07Z","lastTransitionTime":"2026-01-21T17:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.533709 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 16:30:07.126193969 +0000 UTC Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.629913 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.629960 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.629970 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.629989 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.629998 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:07Z","lastTransitionTime":"2026-01-21T17:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.732644 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.732683 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.732692 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.732708 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.732718 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:07Z","lastTransitionTime":"2026-01-21T17:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.835430 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.835512 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.835531 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.835558 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.835579 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:07Z","lastTransitionTime":"2026-01-21T17:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.884528 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.884573 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.884583 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.884603 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.884616 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:07Z","lastTransitionTime":"2026-01-21T17:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:07 crc kubenswrapper[4808]: E0121 17:57:07.902807 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:07Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.908006 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.908041 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.908050 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.908066 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.908096 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:07Z","lastTransitionTime":"2026-01-21T17:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:07 crc kubenswrapper[4808]: E0121 17:57:07.925808 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:07Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.938766 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.938827 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.938841 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.938860 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.939184 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:07Z","lastTransitionTime":"2026-01-21T17:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:07 crc kubenswrapper[4808]: E0121 17:57:07.960131 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:07Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.965383 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.965423 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.965433 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.965449 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.965460 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:07Z","lastTransitionTime":"2026-01-21T17:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:07 crc kubenswrapper[4808]: E0121 17:57:07.981104 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:07Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.985807 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.985842 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.985858 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.985881 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:07 crc kubenswrapper[4808]: I0121 17:57:07.985896 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:07Z","lastTransitionTime":"2026-01-21T17:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:08 crc kubenswrapper[4808]: E0121 17:57:08.001529 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:07Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:08 crc kubenswrapper[4808]: E0121 17:57:08.001697 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.003580 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.003645 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.003665 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.003716 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.003741 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:08Z","lastTransitionTime":"2026-01-21T17:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.106325 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.106367 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.106387 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.106420 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.106444 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:08Z","lastTransitionTime":"2026-01-21T17:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.209017 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.209099 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.209117 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.209152 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.209168 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:08Z","lastTransitionTime":"2026-01-21T17:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.312230 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.312312 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.312349 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.312378 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.312399 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:08Z","lastTransitionTime":"2026-01-21T17:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.415412 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.415474 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.415491 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.415509 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.415522 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:08Z","lastTransitionTime":"2026-01-21T17:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.518807 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.518871 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.518886 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.518909 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.518925 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:08Z","lastTransitionTime":"2026-01-21T17:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.534393 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 23:56:33.699971775 +0000 UTC Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.621418 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.621470 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.621482 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.621500 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.621512 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:08Z","lastTransitionTime":"2026-01-21T17:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.724313 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.724377 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.724390 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.724417 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.724431 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:08Z","lastTransitionTime":"2026-01-21T17:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.828040 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.828101 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.828111 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.828129 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.828141 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:08Z","lastTransitionTime":"2026-01-21T17:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.930901 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.930942 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.930951 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.930969 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:08 crc kubenswrapper[4808]: I0121 17:57:08.930988 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:08Z","lastTransitionTime":"2026-01-21T17:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.033534 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.033607 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.033625 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.033652 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.033669 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:09Z","lastTransitionTime":"2026-01-21T17:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.136616 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.136702 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.136720 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.136780 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.136798 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:09Z","lastTransitionTime":"2026-01-21T17:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.239987 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.240045 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.240062 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.240086 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.240103 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:09Z","lastTransitionTime":"2026-01-21T17:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.342876 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.342911 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.342919 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.342933 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.342942 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:09Z","lastTransitionTime":"2026-01-21T17:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.445060 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.445134 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.445156 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.445186 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.445209 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:09Z","lastTransitionTime":"2026-01-21T17:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.520463 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.520548 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:09 crc kubenswrapper[4808]: E0121 17:57:09.520606 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.520463 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:09 crc kubenswrapper[4808]: E0121 17:57:09.520687 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.520709 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:09 crc kubenswrapper[4808]: E0121 17:57:09.520789 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:09 crc kubenswrapper[4808]: E0121 17:57:09.520876 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.534484 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 08:19:33.593691846 +0000 UTC Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.547459 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.547516 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.547532 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.547558 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.547580 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:09Z","lastTransitionTime":"2026-01-21T17:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.650066 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.650111 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.650124 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.650141 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.650153 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:09Z","lastTransitionTime":"2026-01-21T17:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.753508 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.753566 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.753576 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.753591 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.753601 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:09Z","lastTransitionTime":"2026-01-21T17:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.856479 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.856532 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.856546 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.856569 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.856584 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:09Z","lastTransitionTime":"2026-01-21T17:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.959240 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.959310 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.959321 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.959341 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:09 crc kubenswrapper[4808]: I0121 17:57:09.959354 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:09Z","lastTransitionTime":"2026-01-21T17:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.088767 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.088804 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.088813 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.088827 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.088836 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:10Z","lastTransitionTime":"2026-01-21T17:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.191378 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.191409 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.191416 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.191430 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.191438 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:10Z","lastTransitionTime":"2026-01-21T17:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.295844 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.295915 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.295925 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.295944 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.295963 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:10Z","lastTransitionTime":"2026-01-21T17:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.402824 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.402866 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.402874 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.402888 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.402896 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:10Z","lastTransitionTime":"2026-01-21T17:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.509268 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.509327 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.509342 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.509360 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.509372 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:10Z","lastTransitionTime":"2026-01-21T17:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.535272 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 08:16:08.222783806 +0000 UTC Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.611995 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.612039 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.612048 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.612072 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.612082 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:10Z","lastTransitionTime":"2026-01-21T17:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.715001 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.715043 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.715052 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.715070 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.715080 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:10Z","lastTransitionTime":"2026-01-21T17:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.818211 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.818273 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.818287 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.818308 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.818322 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:10Z","lastTransitionTime":"2026-01-21T17:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.920286 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.920343 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.920367 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.920388 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:10 crc kubenswrapper[4808]: I0121 17:57:10.920403 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:10Z","lastTransitionTime":"2026-01-21T17:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.023319 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.023350 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.023359 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.023374 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.023384 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:11Z","lastTransitionTime":"2026-01-21T17:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.126094 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.126161 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.126177 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.126204 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.126221 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:11Z","lastTransitionTime":"2026-01-21T17:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.228793 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.228847 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.228865 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.228891 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.228908 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:11Z","lastTransitionTime":"2026-01-21T17:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.332483 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.332542 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.332562 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.332589 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.332608 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:11Z","lastTransitionTime":"2026-01-21T17:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.435024 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.435069 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.435081 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.435098 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.435110 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:11Z","lastTransitionTime":"2026-01-21T17:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.520405 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:11 crc kubenswrapper[4808]: E0121 17:57:11.520561 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.520987 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.521311 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:11 crc kubenswrapper[4808]: E0121 17:57:11.521383 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.521577 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:11 crc kubenswrapper[4808]: E0121 17:57:11.521647 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:11 crc kubenswrapper[4808]: E0121 17:57:11.521754 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.521866 4808 scope.go:117] "RemoveContainer" containerID="b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f" Jan 21 17:57:11 crc kubenswrapper[4808]: E0121 17:57:11.523489 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fc2fk_openshift-ovn-kubernetes(bb13a20f-f4ab-412c-8165-5b93edc79628)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.532804 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b148e21a-1611-43f8-85c3-ac3899dde92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65ea8da4528954ea712538a48e6e1c6a1fe893873742ad120f6300ee32f6c95b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59727192447c6eef4b045ac5a9a0c0a7bba9c0f2672d46ff1fc9b2b58c022279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94585d6e08b85c96da3d90042ca20e0a194e7488185ef798b4c2dcc2c2948897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:11Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.535776 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 10:30:55.02308907 +0000 UTC Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.538614 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.538665 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.538687 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.538709 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.538722 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:11Z","lastTransitionTime":"2026-01-21T17:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.546207 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:11Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.567743 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"ble:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:static_routes Mutator:insert Value:{GoSet:[{GoUUID:8944024f-deb7-4076-afb3-4b50a2ff4b4b}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 17:56:56.395979 6462 obj_retry.go:551] Creating *factory.egressNode crc took: 2.120476ms\\\\nI0121 17:56:56.396002 6462 factory.go:1336] Added *v1.Node event handler 7\\\\nI0121 17:56:56.396015 6462 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0121 17:56:56.396027 6462 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0121 17:56:56.396029 6462 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0121 17:56:56.396062 6462 factory.go:656] Stopping watch factory\\\\nI0121 17:56:56.396086 6462 handler.go:208] Removed *v1.Node event handler 7\\\\nI0121 17:56:56.396134 6462 handler.go:208] Removed *v1.Node event handler 2\\\\nI0121 17:56:56.396285 6462 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0121 17:56:56.396360 6462 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0121 17:56:56.396393 6462 ovnkube.go:599] Stopped ovnkube\\\\nI0121 17:56:56.396421 6462 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0121 17:56:56.396480 6462 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fc2fk_openshift-ovn-kubernetes(bb13a20f-f4ab-412c-8165-5b93edc79628)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:11Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.580295 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gm2t2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d39c565-f404-451e-896f-ecc5af76a000\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gm2t2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:11Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.593395 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:11Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.607517 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:11Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.620870 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:11Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.637078 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:11Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.641212 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.641276 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.641287 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.641306 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.641318 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:11Z","lastTransitionTime":"2026-01-21T17:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.647496 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:11Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.659120 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53885bfd-41c1-49a4-beb3-e937534a6ef8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce09f9763152c19c6f0cc141317c798d49f50066d1f0184a114085a19fc9adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90dfe0143cdd3cd582497eddee45eeb42dd96a03eb26a3b753c8aa2fb37436bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw46g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:11Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.670646 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:11Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.682878 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:11Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.698045 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:11Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.709122 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:11Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.724443 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:11Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.740003 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:11Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.743301 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.743334 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.743348 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.743366 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.743377 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:11Z","lastTransitionTime":"2026-01-21T17:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.749608 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:11Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.845766 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.845830 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.845841 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.845858 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.845867 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:11Z","lastTransitionTime":"2026-01-21T17:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.947672 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.947704 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.947712 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.947724 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:11 crc kubenswrapper[4808]: I0121 17:57:11.947732 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:11Z","lastTransitionTime":"2026-01-21T17:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.050176 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.050221 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.050234 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.050267 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.050280 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:12Z","lastTransitionTime":"2026-01-21T17:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.152814 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.152885 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.152910 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.152941 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.152964 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:12Z","lastTransitionTime":"2026-01-21T17:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.255852 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.255894 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.255910 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.255935 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.255953 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:12Z","lastTransitionTime":"2026-01-21T17:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.357956 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.358012 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.358023 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.358037 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.358045 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:12Z","lastTransitionTime":"2026-01-21T17:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.460795 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.460861 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.460885 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.460914 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.460936 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:12Z","lastTransitionTime":"2026-01-21T17:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.536883 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 04:04:31.457244732 +0000 UTC Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.563961 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.564001 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.564013 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.564029 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.564038 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:12Z","lastTransitionTime":"2026-01-21T17:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.666513 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.666554 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.666565 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.666580 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.666590 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:12Z","lastTransitionTime":"2026-01-21T17:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.768888 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.768928 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.768940 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.768955 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.768965 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:12Z","lastTransitionTime":"2026-01-21T17:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.871831 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.871892 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.871909 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.871933 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.871954 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:12Z","lastTransitionTime":"2026-01-21T17:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.974236 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.974299 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.974310 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.974329 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:12 crc kubenswrapper[4808]: I0121 17:57:12.974338 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:12Z","lastTransitionTime":"2026-01-21T17:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.077202 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.077263 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.077281 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.077297 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.077306 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:13Z","lastTransitionTime":"2026-01-21T17:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.181046 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.181284 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.181298 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.181314 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.181325 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:13Z","lastTransitionTime":"2026-01-21T17:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.283560 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.283627 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.283651 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.283680 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.283703 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:13Z","lastTransitionTime":"2026-01-21T17:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.386467 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.386517 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.386527 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.386549 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.386562 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:13Z","lastTransitionTime":"2026-01-21T17:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.489510 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.489585 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.489595 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.489615 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.489628 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:13Z","lastTransitionTime":"2026-01-21T17:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.520385 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.520478 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:13 crc kubenswrapper[4808]: E0121 17:57:13.520551 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.520561 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.520586 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:13 crc kubenswrapper[4808]: E0121 17:57:13.520701 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:13 crc kubenswrapper[4808]: E0121 17:57:13.520909 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:13 crc kubenswrapper[4808]: E0121 17:57:13.521006 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.537805 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 14:15:31.760689276 +0000 UTC Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.592326 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.592377 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.592388 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.592407 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.592418 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:13Z","lastTransitionTime":"2026-01-21T17:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.695043 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.695101 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.695114 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.695134 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.695149 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:13Z","lastTransitionTime":"2026-01-21T17:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.797868 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.797915 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.797926 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.797946 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.797959 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:13Z","lastTransitionTime":"2026-01-21T17:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.900729 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.900772 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.900785 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.900802 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:13 crc kubenswrapper[4808]: I0121 17:57:13.900815 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:13Z","lastTransitionTime":"2026-01-21T17:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.004185 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.004230 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.004264 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.004282 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.004296 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:14Z","lastTransitionTime":"2026-01-21T17:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.107129 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.107178 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.107191 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.107209 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.107220 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:14Z","lastTransitionTime":"2026-01-21T17:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.209772 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.209810 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.209822 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.209839 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.209850 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:14Z","lastTransitionTime":"2026-01-21T17:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.312411 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.312464 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.312481 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.312502 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.312519 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:14Z","lastTransitionTime":"2026-01-21T17:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.415186 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.415283 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.415335 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.415366 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.415384 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:14Z","lastTransitionTime":"2026-01-21T17:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.519084 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.519337 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.519383 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.519410 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.519428 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:14Z","lastTransitionTime":"2026-01-21T17:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.538904 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 04:46:07.092355409 +0000 UTC Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.622119 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.622183 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.622201 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.622229 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.622297 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:14Z","lastTransitionTime":"2026-01-21T17:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.725284 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.725341 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.725353 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.725379 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.725395 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:14Z","lastTransitionTime":"2026-01-21T17:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.828742 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.828833 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.828854 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.828879 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.828892 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:14Z","lastTransitionTime":"2026-01-21T17:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.930422 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.930494 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.930507 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.930524 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:14 crc kubenswrapper[4808]: I0121 17:57:14.930536 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:14Z","lastTransitionTime":"2026-01-21T17:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.032430 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.032485 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.032499 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.032515 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.032527 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:15Z","lastTransitionTime":"2026-01-21T17:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.134324 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.134599 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.134665 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.134731 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.134795 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:15Z","lastTransitionTime":"2026-01-21T17:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.236453 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.236784 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.236933 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.237054 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.237131 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:15Z","lastTransitionTime":"2026-01-21T17:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.339778 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.339815 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.339826 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.339844 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.339858 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:15Z","lastTransitionTime":"2026-01-21T17:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.343762 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs\") pod \"network-metrics-daemon-gm2t2\" (UID: \"8d39c565-f404-451e-896f-ecc5af76a000\") " pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:15 crc kubenswrapper[4808]: E0121 17:57:15.343928 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 17:57:15 crc kubenswrapper[4808]: E0121 17:57:15.343984 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs podName:8d39c565-f404-451e-896f-ecc5af76a000 nodeName:}" failed. No retries permitted until 2026-01-21 17:57:47.343965336 +0000 UTC m=+102.074476221 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs") pod "network-metrics-daemon-gm2t2" (UID: "8d39c565-f404-451e-896f-ecc5af76a000") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.442591 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.442669 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.442684 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.442708 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.442722 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:15Z","lastTransitionTime":"2026-01-21T17:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.520537 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.520538 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.520551 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.520617 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:15 crc kubenswrapper[4808]: E0121 17:57:15.520809 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:15 crc kubenswrapper[4808]: E0121 17:57:15.521040 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:15 crc kubenswrapper[4808]: E0121 17:57:15.521194 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:15 crc kubenswrapper[4808]: E0121 17:57:15.521289 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.539105 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:15Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.539564 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 20:03:05.831072494 +0000 UTC Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.545146 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.545210 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.545229 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.545288 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.545308 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:15Z","lastTransitionTime":"2026-01-21T17:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.560311 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"ble:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:static_routes Mutator:insert Value:{GoSet:[{GoUUID:8944024f-deb7-4076-afb3-4b50a2ff4b4b}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 17:56:56.395979 6462 obj_retry.go:551] Creating *factory.egressNode crc took: 2.120476ms\\\\nI0121 17:56:56.396002 6462 factory.go:1336] Added *v1.Node event handler 7\\\\nI0121 17:56:56.396015 6462 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0121 17:56:56.396027 6462 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0121 17:56:56.396029 6462 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0121 17:56:56.396062 6462 factory.go:656] Stopping watch factory\\\\nI0121 17:56:56.396086 6462 handler.go:208] Removed *v1.Node event handler 7\\\\nI0121 17:56:56.396134 6462 handler.go:208] Removed *v1.Node event handler 2\\\\nI0121 17:56:56.396285 6462 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0121 17:56:56.396360 6462 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0121 17:56:56.396393 6462 ovnkube.go:599] Stopped ovnkube\\\\nI0121 17:56:56.396421 6462 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0121 17:56:56.396480 6462 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fc2fk_openshift-ovn-kubernetes(bb13a20f-f4ab-412c-8165-5b93edc79628)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:15Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.572719 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gm2t2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d39c565-f404-451e-896f-ecc5af76a000\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gm2t2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:15Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.585210 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:15Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.595303 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b148e21a-1611-43f8-85c3-ac3899dde92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65ea8da4528954ea712538a48e6e1c6a1fe893873742ad120f6300ee32f6c95b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59727192447c6eef4b045ac5a9a0c0a7bba9c0f2672d46ff1fc9b2b58c022279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94585d6e08b85c96da3d90042ca20e0a194e7488185ef798b4c2dcc2c2948897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:15Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.606392 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:15Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.619805 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:15Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.630394 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:15Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.642570 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53885bfd-41c1-49a4-beb3-e937534a6ef8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce09f9763152c19c6f0cc141317c798d49f50066d1f0184a114085a19fc9adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90dfe0143cdd3cd582497eddee45eeb42dd96a03eb26a3b753c8aa2fb37436bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw46g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:15Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.647915 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.647949 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.647958 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.647972 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.647981 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:15Z","lastTransitionTime":"2026-01-21T17:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.653534 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:15Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.665886 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:15Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.678324 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:15Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.690446 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:15Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.703065 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:15Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.713739 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:15Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.724141 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:15Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.737055 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:15Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.749561 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.749602 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.749614 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.749631 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.749644 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:15Z","lastTransitionTime":"2026-01-21T17:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.851610 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.851654 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.851665 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.851684 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.851696 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:15Z","lastTransitionTime":"2026-01-21T17:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.954768 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.954806 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.954818 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.954836 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:15 crc kubenswrapper[4808]: I0121 17:57:15.954847 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:15Z","lastTransitionTime":"2026-01-21T17:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.057021 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.057071 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.057083 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.057100 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.057112 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:16Z","lastTransitionTime":"2026-01-21T17:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.160197 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.160264 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.160278 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.160296 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.160307 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:16Z","lastTransitionTime":"2026-01-21T17:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.265687 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.265722 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.265731 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.265757 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.265768 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:16Z","lastTransitionTime":"2026-01-21T17:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.368075 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.368113 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.368123 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.368137 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.368145 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:16Z","lastTransitionTime":"2026-01-21T17:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.470784 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.470816 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.470825 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.470838 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.470847 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:16Z","lastTransitionTime":"2026-01-21T17:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.539678 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 03:06:31.115980147 +0000 UTC Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.572904 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.572946 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.572956 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.572974 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.572985 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:16Z","lastTransitionTime":"2026-01-21T17:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.675759 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.675814 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.675830 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.675858 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.675875 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:16Z","lastTransitionTime":"2026-01-21T17:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.779124 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.779190 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.779199 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.779215 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.779223 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:16Z","lastTransitionTime":"2026-01-21T17:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.881354 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.881383 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.881391 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.881404 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.881412 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:16Z","lastTransitionTime":"2026-01-21T17:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.985597 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.985685 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.985712 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.985745 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:16 crc kubenswrapper[4808]: I0121 17:57:16.985768 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:16Z","lastTransitionTime":"2026-01-21T17:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.088749 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.088817 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.088828 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.088846 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.088858 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:17Z","lastTransitionTime":"2026-01-21T17:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.191099 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.191157 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.191174 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.191196 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.191211 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:17Z","lastTransitionTime":"2026-01-21T17:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.293030 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.293074 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.293083 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.293101 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.293112 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:17Z","lastTransitionTime":"2026-01-21T17:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.395412 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.395482 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.395501 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.395527 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.395545 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:17Z","lastTransitionTime":"2026-01-21T17:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.498038 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.498079 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.498093 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.498109 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.498120 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:17Z","lastTransitionTime":"2026-01-21T17:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.521436 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.521452 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:17 crc kubenswrapper[4808]: E0121 17:57:17.521550 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.521577 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.521587 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:17 crc kubenswrapper[4808]: E0121 17:57:17.521676 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:17 crc kubenswrapper[4808]: E0121 17:57:17.521789 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:17 crc kubenswrapper[4808]: E0121 17:57:17.521854 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.540719 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 11:07:20.464002679 +0000 UTC Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.599916 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.600010 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.600026 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.600046 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.600060 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:17Z","lastTransitionTime":"2026-01-21T17:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.702453 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.702495 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.702505 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.702519 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.702528 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:17Z","lastTransitionTime":"2026-01-21T17:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.805346 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.805398 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.805408 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.805425 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.805438 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:17Z","lastTransitionTime":"2026-01-21T17:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.909018 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.909065 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.909075 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.909091 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.909102 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:17Z","lastTransitionTime":"2026-01-21T17:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.920369 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8fp9_538d1d44-6c67-450a-809c-d9008128ec9f/kube-multus/0.log" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.920429 4808 generic.go:334] "Generic (PLEG): container finished" podID="538d1d44-6c67-450a-809c-d9008128ec9f" containerID="33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397" exitCode=1 Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.920467 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8fp9" event={"ID":"538d1d44-6c67-450a-809c-d9008128ec9f","Type":"ContainerDied","Data":"33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397"} Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.920880 4808 scope.go:117] "RemoveContainer" containerID="33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.935458 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:17Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.953755 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:17Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.964393 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:17Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.974231 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:17Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:17 crc kubenswrapper[4808]: I0121 17:57:17.985330 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53885bfd-41c1-49a4-beb3-e937534a6ef8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce09f9763152c19c6f0cc141317c798d49f50066d1f0184a114085a19fc9adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90dfe0143cdd3cd582497eddee45eeb42dd96a03eb26a3b753c8aa2fb37436bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw46g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:17Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.004453 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:18Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.011638 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.011667 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.011675 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.011696 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.011707 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:18Z","lastTransitionTime":"2026-01-21T17:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.020734 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:57:17Z\\\",\\\"message\\\":\\\"2026-01-21T17:56:32+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e98cf82e-4c33-4303-9923-29d682db9a9a\\\\n2026-01-21T17:56:32+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e98cf82e-4c33-4303-9923-29d682db9a9a to /host/opt/cni/bin/\\\\n2026-01-21T17:56:32Z [verbose] multus-daemon started\\\\n2026-01-21T17:56:32Z [verbose] Readiness Indicator file check\\\\n2026-01-21T17:57:17Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:18Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.033585 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:18Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.041327 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.041354 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.041362 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.041375 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.041385 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:18Z","lastTransitionTime":"2026-01-21T17:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.046448 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:18Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:18 crc kubenswrapper[4808]: E0121 17:57:18.058121 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:18Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.060502 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:18Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.061720 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.061768 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.061780 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.061799 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.061837 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:18Z","lastTransitionTime":"2026-01-21T17:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.071796 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:18Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:18 crc kubenswrapper[4808]: E0121 17:57:18.073011 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:18Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.081690 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.081722 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.081731 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.081745 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.081754 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:18Z","lastTransitionTime":"2026-01-21T17:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.082972 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:18Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:18 crc kubenswrapper[4808]: E0121 17:57:18.099492 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:18Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.104389 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.104423 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.104433 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.104476 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.104486 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:18Z","lastTransitionTime":"2026-01-21T17:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.109735 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:18Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:18 crc kubenswrapper[4808]: E0121 17:57:18.122992 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:18Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.132199 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.132228 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.132235 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.132264 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.132273 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:18Z","lastTransitionTime":"2026-01-21T17:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.142623 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b148e21a-1611-43f8-85c3-ac3899dde92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65ea8da4528954ea712538a48e6e1c6a1fe893873742ad120f6300ee32f6c95b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59727192447c6eef4b045ac5a9a0c0a7bba9c0f2672d46ff1fc9b2b58c022279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94585d6e08b85c96da3d90042ca20e0a194e7488185ef798b4c2dcc2c2948897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:18Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.162027 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:18Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:18 crc kubenswrapper[4808]: E0121 17:57:18.162654 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:18Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:18 crc kubenswrapper[4808]: E0121 17:57:18.162813 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.164397 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.164427 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.164440 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.164458 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.164468 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:18Z","lastTransitionTime":"2026-01-21T17:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.182176 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"ble:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:static_routes Mutator:insert Value:{GoSet:[{GoUUID:8944024f-deb7-4076-afb3-4b50a2ff4b4b}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 17:56:56.395979 6462 obj_retry.go:551] Creating *factory.egressNode crc took: 2.120476ms\\\\nI0121 17:56:56.396002 6462 factory.go:1336] Added *v1.Node event handler 7\\\\nI0121 17:56:56.396015 6462 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0121 17:56:56.396027 6462 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0121 17:56:56.396029 6462 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0121 17:56:56.396062 6462 factory.go:656] Stopping watch factory\\\\nI0121 17:56:56.396086 6462 handler.go:208] Removed *v1.Node event handler 7\\\\nI0121 17:56:56.396134 6462 handler.go:208] Removed *v1.Node event handler 2\\\\nI0121 17:56:56.396285 6462 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0121 17:56:56.396360 6462 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0121 17:56:56.396393 6462 ovnkube.go:599] Stopped ovnkube\\\\nI0121 17:56:56.396421 6462 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0121 17:56:56.396480 6462 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fc2fk_openshift-ovn-kubernetes(bb13a20f-f4ab-412c-8165-5b93edc79628)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:18Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.191845 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gm2t2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d39c565-f404-451e-896f-ecc5af76a000\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gm2t2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:18Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.267021 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.267043 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.267052 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.267068 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.267077 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:18Z","lastTransitionTime":"2026-01-21T17:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.370337 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.370388 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.370405 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.370428 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.370444 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:18Z","lastTransitionTime":"2026-01-21T17:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.472684 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.472761 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.472777 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.472800 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.472849 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:18Z","lastTransitionTime":"2026-01-21T17:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.541335 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 07:39:41.065918904 +0000 UTC Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.575603 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.575673 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.575698 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.575732 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.575756 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:18Z","lastTransitionTime":"2026-01-21T17:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.678399 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.678454 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.678470 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.678495 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.678512 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:18Z","lastTransitionTime":"2026-01-21T17:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.781080 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.781467 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.781480 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.781498 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.781509 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:18Z","lastTransitionTime":"2026-01-21T17:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.884643 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.884717 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.884739 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.884761 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.884781 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:18Z","lastTransitionTime":"2026-01-21T17:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.927799 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8fp9_538d1d44-6c67-450a-809c-d9008128ec9f/kube-multus/0.log" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.927896 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8fp9" event={"ID":"538d1d44-6c67-450a-809c-d9008128ec9f","Type":"ContainerStarted","Data":"f9074967ffe383fc367aaa834087e74f523d9563ecae59d79810bb389f05fbdc"} Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.952694 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:18Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.973460 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b148e21a-1611-43f8-85c3-ac3899dde92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65ea8da4528954ea712538a48e6e1c6a1fe893873742ad120f6300ee32f6c95b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59727192447c6eef4b045ac5a9a0c0a7bba9c0f2672d46ff1fc9b2b58c022279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94585d6e08b85c96da3d90042ca20e0a194e7488185ef798b4c2dcc2c2948897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:18Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.988793 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.988844 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.988860 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.988885 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.988902 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:18Z","lastTransitionTime":"2026-01-21T17:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:18 crc kubenswrapper[4808]: I0121 17:57:18.992192 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:18Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.018728 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"ble:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:static_routes Mutator:insert Value:{GoSet:[{GoUUID:8944024f-deb7-4076-afb3-4b50a2ff4b4b}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 17:56:56.395979 6462 obj_retry.go:551] Creating *factory.egressNode crc took: 2.120476ms\\\\nI0121 17:56:56.396002 6462 factory.go:1336] Added *v1.Node event handler 7\\\\nI0121 17:56:56.396015 6462 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0121 17:56:56.396027 6462 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0121 17:56:56.396029 6462 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0121 17:56:56.396062 6462 factory.go:656] Stopping watch factory\\\\nI0121 17:56:56.396086 6462 handler.go:208] Removed *v1.Node event handler 7\\\\nI0121 17:56:56.396134 6462 handler.go:208] Removed *v1.Node event handler 2\\\\nI0121 17:56:56.396285 6462 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0121 17:56:56.396360 6462 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0121 17:56:56.396393 6462 ovnkube.go:599] Stopped ovnkube\\\\nI0121 17:56:56.396421 6462 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0121 17:56:56.396480 6462 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fc2fk_openshift-ovn-kubernetes(bb13a20f-f4ab-412c-8165-5b93edc79628)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:19Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.034646 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gm2t2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d39c565-f404-451e-896f-ecc5af76a000\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gm2t2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:19Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.051930 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:19Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.068549 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:19Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.084964 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:19Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.094657 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.095187 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.095469 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.095649 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.095665 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:19Z","lastTransitionTime":"2026-01-21T17:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.098651 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:19Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.113090 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53885bfd-41c1-49a4-beb3-e937534a6ef8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce09f9763152c19c6f0cc141317c798d49f50066d1f0184a114085a19fc9adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90dfe0143cdd3cd582497eddee45eeb42dd96a03eb26a3b753c8aa2fb37436bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw46g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:19Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.128917 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9074967ffe383fc367aaa834087e74f523d9563ecae59d79810bb389f05fbdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:57:17Z\\\",\\\"message\\\":\\\"2026-01-21T17:56:32+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e98cf82e-4c33-4303-9923-29d682db9a9a\\\\n2026-01-21T17:56:32+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e98cf82e-4c33-4303-9923-29d682db9a9a to /host/opt/cni/bin/\\\\n2026-01-21T17:56:32Z [verbose] multus-daemon started\\\\n2026-01-21T17:56:32Z [verbose] Readiness Indicator file check\\\\n2026-01-21T17:57:17Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:19Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.141794 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:19Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.157815 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:19Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.173688 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:19Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.189431 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:19Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.198052 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.198144 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.198164 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.198218 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.198288 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:19Z","lastTransitionTime":"2026-01-21T17:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.202610 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:19Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.218068 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:19Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.300841 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.300931 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.300955 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.300984 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.301007 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:19Z","lastTransitionTime":"2026-01-21T17:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.403791 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.403890 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.403907 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.403929 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.403944 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:19Z","lastTransitionTime":"2026-01-21T17:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.506970 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.507008 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.507015 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.507029 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.507039 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:19Z","lastTransitionTime":"2026-01-21T17:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.520407 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.520451 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.520728 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.520965 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:19 crc kubenswrapper[4808]: E0121 17:57:19.521069 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:19 crc kubenswrapper[4808]: E0121 17:57:19.521027 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:19 crc kubenswrapper[4808]: E0121 17:57:19.520974 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:19 crc kubenswrapper[4808]: E0121 17:57:19.521441 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.542250 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 17:19:02.775531768 +0000 UTC Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.609991 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.610063 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.610092 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.610125 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.610151 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:19Z","lastTransitionTime":"2026-01-21T17:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.712488 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.712535 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.712549 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.712565 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.712578 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:19Z","lastTransitionTime":"2026-01-21T17:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.815517 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.815879 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.816042 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.816188 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.816383 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:19Z","lastTransitionTime":"2026-01-21T17:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.919527 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.919586 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.919606 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.919642 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:19 crc kubenswrapper[4808]: I0121 17:57:19.919663 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:19Z","lastTransitionTime":"2026-01-21T17:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.021927 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.021957 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.021965 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.021977 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.021986 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:20Z","lastTransitionTime":"2026-01-21T17:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.125225 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.125295 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.125310 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.125328 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.125341 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:20Z","lastTransitionTime":"2026-01-21T17:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.229885 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.229947 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.229969 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.229999 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.230021 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:20Z","lastTransitionTime":"2026-01-21T17:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.332643 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.332686 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.332698 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.332719 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.332734 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:20Z","lastTransitionTime":"2026-01-21T17:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.435515 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.435565 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.435578 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.435596 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.435608 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:20Z","lastTransitionTime":"2026-01-21T17:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.538553 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.538603 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.538614 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.538633 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.538645 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:20Z","lastTransitionTime":"2026-01-21T17:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.543037 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 21:26:38.359309786 +0000 UTC Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.641379 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.641449 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.641471 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.641499 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.641521 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:20Z","lastTransitionTime":"2026-01-21T17:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.744851 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.744914 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.744939 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.744971 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.744993 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:20Z","lastTransitionTime":"2026-01-21T17:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.848608 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.848685 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.848709 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.848739 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.848761 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:20Z","lastTransitionTime":"2026-01-21T17:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.951186 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.951303 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.951329 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.951357 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:20 crc kubenswrapper[4808]: I0121 17:57:20.951378 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:20Z","lastTransitionTime":"2026-01-21T17:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.055480 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.055567 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.055590 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.055623 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.055643 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:21Z","lastTransitionTime":"2026-01-21T17:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.159028 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.159101 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.159122 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.159150 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.159170 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:21Z","lastTransitionTime":"2026-01-21T17:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.263066 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.263141 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.263166 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.263199 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.263221 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:21Z","lastTransitionTime":"2026-01-21T17:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.365677 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.365904 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.365927 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.365959 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.365976 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:21Z","lastTransitionTime":"2026-01-21T17:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.469657 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.469722 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.469738 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.469762 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.469781 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:21Z","lastTransitionTime":"2026-01-21T17:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.519815 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.519918 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:21 crc kubenswrapper[4808]: E0121 17:57:21.520042 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.520081 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:21 crc kubenswrapper[4808]: E0121 17:57:21.520320 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.520334 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:21 crc kubenswrapper[4808]: E0121 17:57:21.520481 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:21 crc kubenswrapper[4808]: E0121 17:57:21.520645 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.543708 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 02:31:46.594198939 +0000 UTC Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.573411 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.573484 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.573505 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.573549 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.573571 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:21Z","lastTransitionTime":"2026-01-21T17:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.683914 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.684113 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.684166 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.684204 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.684227 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:21Z","lastTransitionTime":"2026-01-21T17:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.787597 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.787676 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.787694 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.787721 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.787739 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:21Z","lastTransitionTime":"2026-01-21T17:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.890876 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.890925 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.890934 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.890951 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.890959 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:21Z","lastTransitionTime":"2026-01-21T17:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.993577 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.993642 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.993659 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.993685 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:21 crc kubenswrapper[4808]: I0121 17:57:21.993703 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:21Z","lastTransitionTime":"2026-01-21T17:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.096944 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.097006 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.097022 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.097051 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.097070 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:22Z","lastTransitionTime":"2026-01-21T17:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.199585 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.199655 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.199679 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.199713 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.199738 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:22Z","lastTransitionTime":"2026-01-21T17:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.302200 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.302237 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.302258 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.302272 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.302281 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:22Z","lastTransitionTime":"2026-01-21T17:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.406194 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.406299 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.406337 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.406370 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.406393 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:22Z","lastTransitionTime":"2026-01-21T17:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.508415 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.508482 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.508492 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.508531 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.508546 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:22Z","lastTransitionTime":"2026-01-21T17:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.544550 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 13:36:11.041013442 +0000 UTC Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.611343 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.611425 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.611450 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.611476 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.611495 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:22Z","lastTransitionTime":"2026-01-21T17:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.714541 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.714602 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.714642 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.714839 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.714877 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:22Z","lastTransitionTime":"2026-01-21T17:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.818871 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.818926 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.818937 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.818955 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.818968 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:22Z","lastTransitionTime":"2026-01-21T17:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.921851 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.921920 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.921942 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.921970 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:22 crc kubenswrapper[4808]: I0121 17:57:22.921992 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:22Z","lastTransitionTime":"2026-01-21T17:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.025536 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.025581 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.025593 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.025610 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.025622 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:23Z","lastTransitionTime":"2026-01-21T17:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.128696 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.128777 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.128796 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.128821 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.128838 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:23Z","lastTransitionTime":"2026-01-21T17:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.231485 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.231568 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.231597 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.231666 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.231693 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:23Z","lastTransitionTime":"2026-01-21T17:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.334730 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.334803 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.334823 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.334847 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.334865 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:23Z","lastTransitionTime":"2026-01-21T17:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.437583 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.437615 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.437624 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.437640 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.437650 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:23Z","lastTransitionTime":"2026-01-21T17:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.520082 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.520150 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.520192 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.520129 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:23 crc kubenswrapper[4808]: E0121 17:57:23.520433 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:23 crc kubenswrapper[4808]: E0121 17:57:23.520570 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:23 crc kubenswrapper[4808]: E0121 17:57:23.520716 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:23 crc kubenswrapper[4808]: E0121 17:57:23.520925 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.540705 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.540779 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.540798 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.540820 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.540840 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:23Z","lastTransitionTime":"2026-01-21T17:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.545017 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 12:17:21.752378658 +0000 UTC Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.644232 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.644316 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.644335 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.644358 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.644377 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:23Z","lastTransitionTime":"2026-01-21T17:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.747547 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.747602 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.747618 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.747641 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.747660 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:23Z","lastTransitionTime":"2026-01-21T17:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.850946 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.851016 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.851040 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.851069 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.851091 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:23Z","lastTransitionTime":"2026-01-21T17:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.953596 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.953660 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.953733 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.953763 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:23 crc kubenswrapper[4808]: I0121 17:57:23.953784 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:23Z","lastTransitionTime":"2026-01-21T17:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.056756 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.056831 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.056848 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.056874 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.056900 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:24Z","lastTransitionTime":"2026-01-21T17:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.161289 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.161335 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.161351 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.161375 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.161392 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:24Z","lastTransitionTime":"2026-01-21T17:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.264755 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.264854 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.264884 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.264958 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.264985 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:24Z","lastTransitionTime":"2026-01-21T17:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.368188 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.368273 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.368297 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.368328 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.368350 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:24Z","lastTransitionTime":"2026-01-21T17:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.471531 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.471591 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.471608 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.471635 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.471653 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:24Z","lastTransitionTime":"2026-01-21T17:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.521937 4808 scope.go:117] "RemoveContainer" containerID="b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.545734 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 19:10:08.746782944 +0000 UTC Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.574551 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.574605 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.574617 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.574637 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.574653 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:24Z","lastTransitionTime":"2026-01-21T17:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.677670 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.677738 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.677761 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.677792 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.677828 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:24Z","lastTransitionTime":"2026-01-21T17:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.781441 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.781510 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.781531 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.781560 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.781674 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:24Z","lastTransitionTime":"2026-01-21T17:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.886329 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.886393 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.886410 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.886436 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.886454 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:24Z","lastTransitionTime":"2026-01-21T17:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.951977 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovnkube-controller/2.log" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.961483 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerStarted","Data":"02b5440e1f9434c7998a819e1e62e4d10ad6887724bac037b63d65847c039f48"} Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.962107 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.977715 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:24Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.988577 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.988622 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.988634 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.988657 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.988671 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:24Z","lastTransitionTime":"2026-01-21T17:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:24 crc kubenswrapper[4808]: I0121 17:57:24.998209 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:24Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.018817 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.035160 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.051380 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53885bfd-41c1-49a4-beb3-e937534a6ef8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce09f9763152c19c6f0cc141317c798d49f50066d1f0184a114085a19fc9adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90dfe0143cdd3cd582497eddee45eeb42dd96a03eb26a3b753c8aa2fb37436bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw46g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.063570 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.086298 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.091491 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.091544 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.091556 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.091576 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.091590 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:25Z","lastTransitionTime":"2026-01-21T17:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.102607 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.120284 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.131873 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.150655 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.171328 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9074967ffe383fc367aaa834087e74f523d9563ecae59d79810bb389f05fbdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:57:17Z\\\",\\\"message\\\":\\\"2026-01-21T17:56:32+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e98cf82e-4c33-4303-9923-29d682db9a9a\\\\n2026-01-21T17:56:32+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e98cf82e-4c33-4303-9923-29d682db9a9a to /host/opt/cni/bin/\\\\n2026-01-21T17:56:32Z [verbose] multus-daemon started\\\\n2026-01-21T17:56:32Z [verbose] Readiness Indicator file check\\\\n2026-01-21T17:57:17Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.192309 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.194209 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.194274 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.194287 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.194311 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.194325 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:25Z","lastTransitionTime":"2026-01-21T17:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.213876 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b148e21a-1611-43f8-85c3-ac3899dde92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65ea8da4528954ea712538a48e6e1c6a1fe893873742ad120f6300ee32f6c95b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59727192447c6eef4b045ac5a9a0c0a7bba9c0f2672d46ff1fc9b2b58c022279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94585d6e08b85c96da3d90042ca20e0a194e7488185ef798b4c2dcc2c2948897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.234360 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.254517 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02b5440e1f9434c7998a819e1e62e4d10ad6887724bac037b63d65847c039f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"ble:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:static_routes Mutator:insert Value:{GoSet:[{GoUUID:8944024f-deb7-4076-afb3-4b50a2ff4b4b}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 17:56:56.395979 6462 obj_retry.go:551] Creating *factory.egressNode crc took: 2.120476ms\\\\nI0121 17:56:56.396002 6462 factory.go:1336] Added *v1.Node event handler 7\\\\nI0121 17:56:56.396015 6462 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0121 17:56:56.396027 6462 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0121 17:56:56.396029 6462 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0121 17:56:56.396062 6462 factory.go:656] Stopping watch factory\\\\nI0121 17:56:56.396086 6462 handler.go:208] Removed *v1.Node event handler 7\\\\nI0121 17:56:56.396134 6462 handler.go:208] Removed *v1.Node event handler 2\\\\nI0121 17:56:56.396285 6462 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0121 17:56:56.396360 6462 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0121 17:56:56.396393 6462 ovnkube.go:599] Stopped ovnkube\\\\nI0121 17:56:56.396421 6462 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0121 17:56:56.396480 6462 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.269432 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gm2t2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d39c565-f404-451e-896f-ecc5af76a000\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gm2t2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.297058 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.297114 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.297128 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.297151 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.297165 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:25Z","lastTransitionTime":"2026-01-21T17:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.400580 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.400632 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.400642 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.400658 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.400672 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:25Z","lastTransitionTime":"2026-01-21T17:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.504078 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.504138 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.504149 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.504309 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.504323 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:25Z","lastTransitionTime":"2026-01-21T17:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.520378 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.520490 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.520486 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:25 crc kubenswrapper[4808]: E0121 17:57:25.520715 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.520761 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:25 crc kubenswrapper[4808]: E0121 17:57:25.520814 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:25 crc kubenswrapper[4808]: E0121 17:57:25.521290 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:25 crc kubenswrapper[4808]: E0121 17:57:25.521493 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.541227 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.545949 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 13:40:17.316720774 +0000 UTC Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.561474 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.580367 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.599562 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.606509 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.606559 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.606572 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.606593 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.606608 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:25Z","lastTransitionTime":"2026-01-21T17:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.618596 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53885bfd-41c1-49a4-beb3-e937534a6ef8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce09f9763152c19c6f0cc141317c798d49f50066d1f0184a114085a19fc9adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90dfe0143cdd3cd582497eddee45eeb42dd96a03eb26a3b753c8aa2fb37436bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw46g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.636880 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.651131 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.665079 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.678092 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.690453 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.709030 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.709090 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.709101 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.709126 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.709140 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:25Z","lastTransitionTime":"2026-01-21T17:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.712214 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.733422 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9074967ffe383fc367aaa834087e74f523d9563ecae59d79810bb389f05fbdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:57:17Z\\\",\\\"message\\\":\\\"2026-01-21T17:56:32+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e98cf82e-4c33-4303-9923-29d682db9a9a\\\\n2026-01-21T17:56:32+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e98cf82e-4c33-4303-9923-29d682db9a9a to /host/opt/cni/bin/\\\\n2026-01-21T17:56:32Z [verbose] multus-daemon started\\\\n2026-01-21T17:56:32Z [verbose] Readiness Indicator file check\\\\n2026-01-21T17:57:17Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.758984 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.786383 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b148e21a-1611-43f8-85c3-ac3899dde92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65ea8da4528954ea712538a48e6e1c6a1fe893873742ad120f6300ee32f6c95b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59727192447c6eef4b045ac5a9a0c0a7bba9c0f2672d46ff1fc9b2b58c022279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94585d6e08b85c96da3d90042ca20e0a194e7488185ef798b4c2dcc2c2948897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.804646 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.812213 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.812281 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.812292 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.812314 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.812348 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:25Z","lastTransitionTime":"2026-01-21T17:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.832362 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02b5440e1f9434c7998a819e1e62e4d10ad6887724bac037b63d65847c039f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"ble:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:static_routes Mutator:insert Value:{GoSet:[{GoUUID:8944024f-deb7-4076-afb3-4b50a2ff4b4b}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 17:56:56.395979 6462 obj_retry.go:551] Creating *factory.egressNode crc took: 2.120476ms\\\\nI0121 17:56:56.396002 6462 factory.go:1336] Added *v1.Node event handler 7\\\\nI0121 17:56:56.396015 6462 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0121 17:56:56.396027 6462 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0121 17:56:56.396029 6462 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0121 17:56:56.396062 6462 factory.go:656] Stopping watch factory\\\\nI0121 17:56:56.396086 6462 handler.go:208] Removed *v1.Node event handler 7\\\\nI0121 17:56:56.396134 6462 handler.go:208] Removed *v1.Node event handler 2\\\\nI0121 17:56:56.396285 6462 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0121 17:56:56.396360 6462 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0121 17:56:56.396393 6462 ovnkube.go:599] Stopped ovnkube\\\\nI0121 17:56:56.396421 6462 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0121 17:56:56.396480 6462 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.847430 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gm2t2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d39c565-f404-451e-896f-ecc5af76a000\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gm2t2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:25Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.916347 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.916427 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.916451 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.916487 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:25 crc kubenswrapper[4808]: I0121 17:57:25.916520 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:25Z","lastTransitionTime":"2026-01-21T17:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.019744 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.019814 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.019832 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.019856 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.019873 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:26Z","lastTransitionTime":"2026-01-21T17:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.123063 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.123109 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.123120 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.123141 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.123155 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:26Z","lastTransitionTime":"2026-01-21T17:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.227419 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.227524 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.227549 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.227581 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.227601 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:26Z","lastTransitionTime":"2026-01-21T17:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.330390 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.330446 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.330463 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.330487 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.330504 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:26Z","lastTransitionTime":"2026-01-21T17:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.434078 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.434163 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.434190 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.434225 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.434283 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:26Z","lastTransitionTime":"2026-01-21T17:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.537642 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.537710 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.537727 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.537752 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.537769 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:26Z","lastTransitionTime":"2026-01-21T17:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.546963 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 23:25:13.916560904 +0000 UTC Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.640443 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.640486 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.640495 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.640510 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.640520 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:26Z","lastTransitionTime":"2026-01-21T17:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.743743 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.743794 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.743806 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.743825 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.743839 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:26Z","lastTransitionTime":"2026-01-21T17:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.847426 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.847496 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.847513 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.847538 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.847554 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:26Z","lastTransitionTime":"2026-01-21T17:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.950972 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.951263 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.951402 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.951535 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.951649 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:26Z","lastTransitionTime":"2026-01-21T17:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.969403 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovnkube-controller/3.log" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.970862 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovnkube-controller/2.log" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.975623 4808 generic.go:334] "Generic (PLEG): container finished" podID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerID="02b5440e1f9434c7998a819e1e62e4d10ad6887724bac037b63d65847c039f48" exitCode=1 Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.975763 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerDied","Data":"02b5440e1f9434c7998a819e1e62e4d10ad6887724bac037b63d65847c039f48"} Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.975886 4808 scope.go:117] "RemoveContainer" containerID="b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.977063 4808 scope.go:117] "RemoveContainer" containerID="02b5440e1f9434c7998a819e1e62e4d10ad6887724bac037b63d65847c039f48" Jan 21 17:57:26 crc kubenswrapper[4808]: E0121 17:57:26.977415 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fc2fk_openshift-ovn-kubernetes(bb13a20f-f4ab-412c-8165-5b93edc79628)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" Jan 21 17:57:26 crc kubenswrapper[4808]: I0121 17:57:26.997926 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:26Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.017893 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.036730 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d76d1c92-28d2-4476-8af9-d95cc779425e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c69d08a8a4cc4186bad71b24da83d587fdaf171c270a8d94244ac0b1d1975c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xp5bs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lgtv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.055163 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.055224 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.055264 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.055293 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.055311 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:27Z","lastTransitionTime":"2026-01-21T17:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.058081 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-sm4d2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6c02096-5729-4118-9b7a-0b3722acf2b6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14962350c1d2e66958f81064d3479c4a50763cb19fa032335ce5c15f860a4ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kx7ms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-sm4d2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.079179 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53885bfd-41c1-49a4-beb3-e937534a6ef8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce09f9763152c19c6f0cc141317c798d49f50066d1f0184a114085a19fc9adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90dfe0143cdd3cd582497eddee45eeb42dd96a03eb26a3b753c8aa2fb37436bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tn7ts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw46g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.105937 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8fp9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"538d1d44-6c67-450a-809c-d9008128ec9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9074967ffe383fc367aaa834087e74f523d9563ecae59d79810bb389f05fbdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:57:17Z\\\",\\\"message\\\":\\\"2026-01-21T17:56:32+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e98cf82e-4c33-4303-9923-29d682db9a9a\\\\n2026-01-21T17:56:32+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e98cf82e-4c33-4303-9923-29d682db9a9a to /host/opt/cni/bin/\\\\n2026-01-21T17:56:32Z [verbose] multus-daemon started\\\\n2026-01-21T17:56:32Z [verbose] Readiness Indicator file check\\\\n2026-01-21T17:57:17Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnpck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8fp9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.120885 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-28p27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b4de769-807e-43ca-b0e6-c3248bb4a6c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ccc2cc48900ac18e770148e9898995a071d39d79741fa227e87786829e7521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-prdf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-28p27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.137490 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dc09b23-0283-4a2d-9a0b-cda4bfa5fcda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca7aa17b7a764339974f1c0e38b15a8405a1cb99e0fae9d776bf820b25ca141a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22463cdb9c8ebc344656ceaeed27c6ee25180a65be5c351aebc0561310c6a4fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6067bb431ef06692499d96ed9bb78447d812bdf859e50080447fe7b4ecbc7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.154439 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327a0f56c7b26551a421f20a470d2abe707ddb2df387ab3efb55b09be10092c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.159439 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.159500 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.159514 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.159555 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.159570 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:27Z","lastTransitionTime":"2026-01-21T17:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.171313 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.187091 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5538dd6d9b0c0cb6ea3435eb784609509b9cc9fe235a149475c1f74de4436bbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.206161 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4kc58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ffed385-f2d1-4043-9323-561451d85fad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710f96a9b4ef5d984d0215f88dc8eed85300eeed7ca2891a765fb2d8ec1f7f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc05c82930893a3a420da192c8eaa86a743d89730a2e6ed5b10b8cd71d92970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759576d0089f98d1e8e55ddbd09e1ff9b2b8a75fa8517e850487219bb567ed17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f1f5e02ee241b09106a8b1965f9f593b51cfa84de6db0d191f598bf2dad8169\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9a7f399b885948c6b91cb6a3a732a3d5067e70a46e78f02e558624b1a33142e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6ff1fc21ef7ec4db361e7fcebc3d91606dfbbbb1304d50462e851ea717040cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2a4589fa83327318d637043b06b2d67f2c7ac9a251bc8950e094c854e98f84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2824\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4kc58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.223396 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bac69ca-5e64-4274-b84e-5fcb32a3e63b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T17:56:23Z\\\",\\\"message\\\":\\\"_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.484269 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 17:56:23.484317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 17:56:23.484345 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 17:56:23.484354 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 17:56:23.489604 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0121 17:56:23.489625 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0121 17:56:23.489633 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489639 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 17:56:23.489643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 17:56:23.489646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 17:56:23.489649 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 17:56:23.489652 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 17:56:23.492917 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492937 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0121 17:56:23.492984 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0121 17:56:23.492987 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nF0121 17:56:23.493028 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.244416 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b148e21a-1611-43f8-85c3-ac3899dde92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65ea8da4528954ea712538a48e6e1c6a1fe893873742ad120f6300ee32f6c95b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59727192447c6eef4b045ac5a9a0c0a7bba9c0f2672d46ff1fc9b2b58c022279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94585d6e08b85c96da3d90042ca20e0a194e7488185ef798b4c2dcc2c2948897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://636015c2a5eed3b18a73c3ec70062016184b634856c394a9a98bafac907705f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.261553 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053a1ee9165ff948a2e3b508e8666d84013ae1b6f9fe87153c057fc41dd3bcc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://003ffc2862e1c5b76912501a0dfc3d88f8dd10d8770ac7d9f2262bf330ee5766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.262683 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.262760 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.262782 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.262814 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.262836 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:27Z","lastTransitionTime":"2026-01-21T17:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.294621 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb13a20f-f4ab-412c-8165-5b93edc79628\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02b5440e1f9434c7998a819e1e62e4d10ad6887724bac037b63d65847c039f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8656cb88d6d725b0c14806f5c5a640337ebd4967346d9755271a8292f2e2b0f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:56:57Z\\\",\\\"message\\\":\\\"ble:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:static_routes Mutator:insert Value:{GoSet:[{GoUUID:8944024f-deb7-4076-afb3-4b50a2ff4b4b}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 17:56:56.395979 6462 obj_retry.go:551] Creating *factory.egressNode crc took: 2.120476ms\\\\nI0121 17:56:56.396002 6462 factory.go:1336] Added *v1.Node event handler 7\\\\nI0121 17:56:56.396015 6462 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0121 17:56:56.396027 6462 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0121 17:56:56.396029 6462 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0121 17:56:56.396062 6462 factory.go:656] Stopping watch factory\\\\nI0121 17:56:56.396086 6462 handler.go:208] Removed *v1.Node event handler 7\\\\nI0121 17:56:56.396134 6462 handler.go:208] Removed *v1.Node event handler 2\\\\nI0121 17:56:56.396285 6462 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0121 17:56:56.396360 6462 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0121 17:56:56.396393 6462 ovnkube.go:599] Stopped ovnkube\\\\nI0121 17:56:56.396421 6462 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0121 17:56:56.396480 6462 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02b5440e1f9434c7998a819e1e62e4d10ad6887724bac037b63d65847c039f48\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T17:57:26Z\\\",\\\"message\\\":\\\"andler 7 for removal\\\\nI0121 17:57:25.630806 6854 handler.go:208] Removed *v1.Node event handler 2\\\\nI0121 17:57:25.630825 6854 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0121 17:57:25.630857 6854 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0121 17:57:25.630878 6854 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0121 17:57:25.630908 6854 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0121 17:57:25.630918 6854 handler.go:208] Removed *v1.Node event handler 7\\\\nI0121 17:57:25.630909 6854 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0121 17:57:25.630941 6854 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0121 17:57:25.630954 6854 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0121 17:57:25.630993 6854 factory.go:656] Stopping watch factory\\\\nI0121 17:57:25.631005 6854 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0121 17:57:25.631040 6854 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0121 17:57:25.631133 6854 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0121 17:57:25.631183 6854 ovnkube.go:599] Stopped ovnkube\\\\nI0121 17:57:25.631223 6854 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0121 17:57:25.631345 6854 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T17:57:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T17:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T17:56:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T17:56:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwp7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc2fk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.306392 4808 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gm2t2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d39c565-f404-451e-896f-ecc5af76a000\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T17:56:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T17:56:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gm2t2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:27Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.365563 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.365619 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.365637 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.365663 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.365683 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:27Z","lastTransitionTime":"2026-01-21T17:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.468710 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.468766 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.468784 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.468808 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.468825 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:27Z","lastTransitionTime":"2026-01-21T17:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.475291 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:57:27 crc kubenswrapper[4808]: E0121 17:57:27.475525 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:31.475501445 +0000 UTC m=+146.206012370 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.520449 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.520507 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.520966 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.520980 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:27 crc kubenswrapper[4808]: E0121 17:57:27.521346 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:27 crc kubenswrapper[4808]: E0121 17:57:27.521903 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:27 crc kubenswrapper[4808]: E0121 17:57:27.522080 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:27 crc kubenswrapper[4808]: E0121 17:57:27.522173 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.539004 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.540157 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.547829 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 14:43:29.865577578 +0000 UTC Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.572080 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.572141 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.572158 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.572232 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.572289 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:27Z","lastTransitionTime":"2026-01-21T17:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.576278 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.576337 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.576392 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.576432 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:27 crc kubenswrapper[4808]: E0121 17:57:27.576553 4808 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 17:57:27 crc kubenswrapper[4808]: E0121 17:57:27.576576 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 17:57:27 crc kubenswrapper[4808]: E0121 17:57:27.576605 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 17:57:27 crc kubenswrapper[4808]: E0121 17:57:27.576625 4808 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:57:27 crc kubenswrapper[4808]: E0121 17:57:27.576651 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 17:58:31.576618241 +0000 UTC m=+146.307129166 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 17:57:27 crc kubenswrapper[4808]: E0121 17:57:27.576681 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 17:58:31.576662922 +0000 UTC m=+146.307173847 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:57:27 crc kubenswrapper[4808]: E0121 17:57:27.576738 4808 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 17:57:27 crc kubenswrapper[4808]: E0121 17:57:27.576763 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 17:57:27 crc kubenswrapper[4808]: E0121 17:57:27.576826 4808 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 17:57:27 crc kubenswrapper[4808]: E0121 17:57:27.576852 4808 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:57:27 crc kubenswrapper[4808]: E0121 17:57:27.576797 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 17:58:31.576780915 +0000 UTC m=+146.307291830 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 17:57:27 crc kubenswrapper[4808]: E0121 17:57:27.577039 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 17:58:31.576946528 +0000 UTC m=+146.307457453 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.675724 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.675779 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.675803 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.675953 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.676029 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:27Z","lastTransitionTime":"2026-01-21T17:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.779724 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.779793 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.779818 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.779849 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.779869 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:27Z","lastTransitionTime":"2026-01-21T17:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.882640 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.882685 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.882699 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.882716 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.882727 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:27Z","lastTransitionTime":"2026-01-21T17:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.981104 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovnkube-controller/3.log" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.984377 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.984474 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.984493 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.984518 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:27 crc kubenswrapper[4808]: I0121 17:57:27.984536 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:27Z","lastTransitionTime":"2026-01-21T17:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.088512 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.088592 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.088611 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.088638 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.088655 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:28Z","lastTransitionTime":"2026-01-21T17:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.191870 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.192045 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.192219 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.192290 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.192316 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:28Z","lastTransitionTime":"2026-01-21T17:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.295982 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.296052 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.296078 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.296108 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.296129 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:28Z","lastTransitionTime":"2026-01-21T17:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.399455 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.399531 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.399553 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.399584 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.399606 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:28Z","lastTransitionTime":"2026-01-21T17:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.502880 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.502941 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.502956 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.502975 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.502988 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:28Z","lastTransitionTime":"2026-01-21T17:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.524999 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.525074 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.525103 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.525137 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.525159 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:28Z","lastTransitionTime":"2026-01-21T17:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:28 crc kubenswrapper[4808]: E0121 17:57:28.545163 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:28Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.547985 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 15:47:26.424279585 +0000 UTC Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.550728 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.550786 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.550798 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.550818 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.550831 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:28Z","lastTransitionTime":"2026-01-21T17:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:28 crc kubenswrapper[4808]: E0121 17:57:28.567078 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:28Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.571292 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.571334 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.571344 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.571362 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.571372 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:28Z","lastTransitionTime":"2026-01-21T17:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:28 crc kubenswrapper[4808]: E0121 17:57:28.593645 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T17:57:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7acf3d6e-f41a-4724-b064-e0293f86b6ec\\\",\\\"systemUUID\\\":\\\"91b163f8-1521-490f-afb6-c160deb5beca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T17:57:28Z is after 2025-08-24T17:21:41Z" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.597898 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.597941 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.597957 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.597977 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.597992 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:28Z","lastTransitionTime":"2026-01-21T17:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.764212 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.764337 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.764364 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.764392 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.764415 4808 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T17:57:28Z","lastTransitionTime":"2026-01-21T17:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.802095 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp"] Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.802684 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.813497 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.813588 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.813666 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.813866 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.830877 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=1.830704315 podStartE2EDuration="1.830704315s" podCreationTimestamp="2026-01-21 17:57:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:57:28.830338766 +0000 UTC m=+83.560849651" watchObservedRunningTime="2026-01-21 17:57:28.830704315 +0000 UTC m=+83.561215200" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.845469 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-sm4d2" podStartSLOduration=59.845428615 podStartE2EDuration="59.845428615s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:57:28.845411404 +0000 UTC m=+83.575922319" watchObservedRunningTime="2026-01-21 17:57:28.845428615 +0000 UTC m=+83.575939500" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.863515 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw46g" podStartSLOduration=59.863481702 podStartE2EDuration="59.863481702s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:57:28.861455795 +0000 UTC m=+83.591966690" watchObservedRunningTime="2026-01-21 17:57:28.863481702 +0000 UTC m=+83.593992627" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.890491 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/36224127-9452-478d-a6d0-20eade410b21-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fh4vp\" (UID: \"36224127-9452-478d-a6d0-20eade410b21\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.890696 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/36224127-9452-478d-a6d0-20eade410b21-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fh4vp\" (UID: \"36224127-9452-478d-a6d0-20eade410b21\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.890751 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36224127-9452-478d-a6d0-20eade410b21-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fh4vp\" (UID: \"36224127-9452-478d-a6d0-20eade410b21\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.890842 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36224127-9452-478d-a6d0-20eade410b21-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fh4vp\" (UID: \"36224127-9452-478d-a6d0-20eade410b21\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.890952 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/36224127-9452-478d-a6d0-20eade410b21-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fh4vp\" (UID: \"36224127-9452-478d-a6d0-20eade410b21\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.919262 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=1.919210649 podStartE2EDuration="1.919210649s" podCreationTimestamp="2026-01-21 17:57:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:57:28.905667407 +0000 UTC m=+83.636178302" watchObservedRunningTime="2026-01-21 17:57:28.919210649 +0000 UTC m=+83.649721544" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.943379 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=60.943342687 podStartE2EDuration="1m0.943342687s" podCreationTimestamp="2026-01-21 17:56:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:57:28.919986577 +0000 UTC m=+83.650497552" watchObservedRunningTime="2026-01-21 17:57:28.943342687 +0000 UTC m=+83.673853612" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.992190 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/36224127-9452-478d-a6d0-20eade410b21-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fh4vp\" (UID: \"36224127-9452-478d-a6d0-20eade410b21\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.992287 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/36224127-9452-478d-a6d0-20eade410b21-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fh4vp\" (UID: \"36224127-9452-478d-a6d0-20eade410b21\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.992384 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/36224127-9452-478d-a6d0-20eade410b21-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fh4vp\" (UID: \"36224127-9452-478d-a6d0-20eade410b21\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.992431 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36224127-9452-478d-a6d0-20eade410b21-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fh4vp\" (UID: \"36224127-9452-478d-a6d0-20eade410b21\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.992515 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36224127-9452-478d-a6d0-20eade410b21-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fh4vp\" (UID: \"36224127-9452-478d-a6d0-20eade410b21\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.992541 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/36224127-9452-478d-a6d0-20eade410b21-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fh4vp\" (UID: \"36224127-9452-478d-a6d0-20eade410b21\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.992376 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/36224127-9452-478d-a6d0-20eade410b21-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fh4vp\" (UID: \"36224127-9452-478d-a6d0-20eade410b21\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.993642 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/36224127-9452-478d-a6d0-20eade410b21-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fh4vp\" (UID: \"36224127-9452-478d-a6d0-20eade410b21\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp" Jan 21 17:57:28 crc kubenswrapper[4808]: I0121 17:57:28.994535 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-4kc58" podStartSLOduration=59.99452261 podStartE2EDuration="59.99452261s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:57:28.994213362 +0000 UTC m=+83.724724287" watchObservedRunningTime="2026-01-21 17:57:28.99452261 +0000 UTC m=+83.725033505" Jan 21 17:57:29 crc kubenswrapper[4808]: I0121 17:57:29.000277 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36224127-9452-478d-a6d0-20eade410b21-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fh4vp\" (UID: \"36224127-9452-478d-a6d0-20eade410b21\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp" Jan 21 17:57:29 crc kubenswrapper[4808]: I0121 17:57:29.011373 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36224127-9452-478d-a6d0-20eade410b21-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fh4vp\" (UID: \"36224127-9452-478d-a6d0-20eade410b21\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp" Jan 21 17:57:29 crc kubenswrapper[4808]: I0121 17:57:29.012267 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-z8fp9" podStartSLOduration=60.012227299 podStartE2EDuration="1m0.012227299s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:57:29.011927442 +0000 UTC m=+83.742438337" watchObservedRunningTime="2026-01-21 17:57:29.012227299 +0000 UTC m=+83.742738194" Jan 21 17:57:29 crc kubenswrapper[4808]: I0121 17:57:29.026440 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-28p27" podStartSLOduration=60.026424307 podStartE2EDuration="1m0.026424307s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:57:29.025348162 +0000 UTC m=+83.755859037" watchObservedRunningTime="2026-01-21 17:57:29.026424307 +0000 UTC m=+83.756935192" Jan 21 17:57:29 crc kubenswrapper[4808]: I0121 17:57:29.084792 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=66.084765355 podStartE2EDuration="1m6.084765355s" podCreationTimestamp="2026-01-21 17:56:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:57:29.067487985 +0000 UTC m=+83.797998890" watchObservedRunningTime="2026-01-21 17:57:29.084765355 +0000 UTC m=+83.815276260" Jan 21 17:57:29 crc kubenswrapper[4808]: I0121 17:57:29.085623 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=38.085617014 podStartE2EDuration="38.085617014s" podCreationTimestamp="2026-01-21 17:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:57:29.08456272 +0000 UTC m=+83.815073625" watchObservedRunningTime="2026-01-21 17:57:29.085617014 +0000 UTC m=+83.816127909" Jan 21 17:57:29 crc kubenswrapper[4808]: I0121 17:57:29.128600 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp" Jan 21 17:57:29 crc kubenswrapper[4808]: I0121 17:57:29.520774 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:29 crc kubenswrapper[4808]: I0121 17:57:29.520851 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:29 crc kubenswrapper[4808]: I0121 17:57:29.520774 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:29 crc kubenswrapper[4808]: E0121 17:57:29.520938 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:29 crc kubenswrapper[4808]: E0121 17:57:29.521079 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:29 crc kubenswrapper[4808]: E0121 17:57:29.521283 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:29 crc kubenswrapper[4808]: I0121 17:57:29.521346 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:29 crc kubenswrapper[4808]: E0121 17:57:29.521433 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:29 crc kubenswrapper[4808]: I0121 17:57:29.548587 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 05:59:03.149137325 +0000 UTC Jan 21 17:57:29 crc kubenswrapper[4808]: I0121 17:57:29.548675 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Jan 21 17:57:29 crc kubenswrapper[4808]: I0121 17:57:29.560412 4808 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 21 17:57:29 crc kubenswrapper[4808]: I0121 17:57:29.996183 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp" event={"ID":"36224127-9452-478d-a6d0-20eade410b21","Type":"ContainerStarted","Data":"82197bfa00af7058ccb2987283f66866090e9844ba2341bfed42e7b724c1c9d2"} Jan 21 17:57:29 crc kubenswrapper[4808]: I0121 17:57:29.996292 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp" event={"ID":"36224127-9452-478d-a6d0-20eade410b21","Type":"ContainerStarted","Data":"aa68f4aa44729d5d3402a32ce3f4febc15d38fd13489ff114e840e8077141cd3"} Jan 21 17:57:30 crc kubenswrapper[4808]: I0121 17:57:30.020713 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podStartSLOduration=61.020677138 podStartE2EDuration="1m1.020677138s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:57:29.202558056 +0000 UTC m=+83.933068941" watchObservedRunningTime="2026-01-21 17:57:30.020677138 +0000 UTC m=+84.751188073" Jan 21 17:57:30 crc kubenswrapper[4808]: I0121 17:57:30.021162 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh4vp" podStartSLOduration=61.021152089 podStartE2EDuration="1m1.021152089s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:57:30.020104485 +0000 UTC m=+84.750615410" watchObservedRunningTime="2026-01-21 17:57:30.021152089 +0000 UTC m=+84.751663014" Jan 21 17:57:31 crc kubenswrapper[4808]: I0121 17:57:31.520214 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:31 crc kubenswrapper[4808]: I0121 17:57:31.520317 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:31 crc kubenswrapper[4808]: I0121 17:57:31.520197 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:31 crc kubenswrapper[4808]: E0121 17:57:31.520444 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:31 crc kubenswrapper[4808]: I0121 17:57:31.520503 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:31 crc kubenswrapper[4808]: E0121 17:57:31.520576 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:31 crc kubenswrapper[4808]: E0121 17:57:31.520695 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:31 crc kubenswrapper[4808]: E0121 17:57:31.520870 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:33 crc kubenswrapper[4808]: I0121 17:57:33.520133 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:33 crc kubenswrapper[4808]: I0121 17:57:33.520286 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:33 crc kubenswrapper[4808]: I0121 17:57:33.520479 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:33 crc kubenswrapper[4808]: E0121 17:57:33.520526 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:33 crc kubenswrapper[4808]: I0121 17:57:33.520608 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:33 crc kubenswrapper[4808]: E0121 17:57:33.520646 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:33 crc kubenswrapper[4808]: E0121 17:57:33.520753 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:33 crc kubenswrapper[4808]: E0121 17:57:33.520798 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:35 crc kubenswrapper[4808]: I0121 17:57:35.520089 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:35 crc kubenswrapper[4808]: I0121 17:57:35.520145 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:35 crc kubenswrapper[4808]: I0121 17:57:35.520114 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:35 crc kubenswrapper[4808]: E0121 17:57:35.521960 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:35 crc kubenswrapper[4808]: I0121 17:57:35.521984 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:35 crc kubenswrapper[4808]: E0121 17:57:35.522062 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:35 crc kubenswrapper[4808]: E0121 17:57:35.522132 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:35 crc kubenswrapper[4808]: E0121 17:57:35.522221 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:37 crc kubenswrapper[4808]: I0121 17:57:37.519871 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:37 crc kubenswrapper[4808]: I0121 17:57:37.519909 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:37 crc kubenswrapper[4808]: I0121 17:57:37.519873 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:37 crc kubenswrapper[4808]: E0121 17:57:37.520214 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:37 crc kubenswrapper[4808]: E0121 17:57:37.520359 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:37 crc kubenswrapper[4808]: I0121 17:57:37.520435 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:37 crc kubenswrapper[4808]: E0121 17:57:37.520482 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:37 crc kubenswrapper[4808]: E0121 17:57:37.520589 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:39 crc kubenswrapper[4808]: I0121 17:57:39.520301 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:39 crc kubenswrapper[4808]: E0121 17:57:39.520472 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:39 crc kubenswrapper[4808]: I0121 17:57:39.520658 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:39 crc kubenswrapper[4808]: I0121 17:57:39.520673 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:39 crc kubenswrapper[4808]: E0121 17:57:39.520996 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:39 crc kubenswrapper[4808]: I0121 17:57:39.520902 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:39 crc kubenswrapper[4808]: E0121 17:57:39.521189 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:39 crc kubenswrapper[4808]: E0121 17:57:39.520971 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:41 crc kubenswrapper[4808]: I0121 17:57:41.520648 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:41 crc kubenswrapper[4808]: I0121 17:57:41.520703 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:41 crc kubenswrapper[4808]: I0121 17:57:41.520776 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:41 crc kubenswrapper[4808]: I0121 17:57:41.520661 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:41 crc kubenswrapper[4808]: E0121 17:57:41.520964 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:41 crc kubenswrapper[4808]: E0121 17:57:41.521095 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:41 crc kubenswrapper[4808]: E0121 17:57:41.521240 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:41 crc kubenswrapper[4808]: E0121 17:57:41.521419 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:42 crc kubenswrapper[4808]: I0121 17:57:42.521378 4808 scope.go:117] "RemoveContainer" containerID="02b5440e1f9434c7998a819e1e62e4d10ad6887724bac037b63d65847c039f48" Jan 21 17:57:42 crc kubenswrapper[4808]: E0121 17:57:42.521645 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fc2fk_openshift-ovn-kubernetes(bb13a20f-f4ab-412c-8165-5b93edc79628)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" Jan 21 17:57:43 crc kubenswrapper[4808]: I0121 17:57:43.520024 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:43 crc kubenswrapper[4808]: I0121 17:57:43.520149 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:43 crc kubenswrapper[4808]: I0121 17:57:43.520659 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:43 crc kubenswrapper[4808]: I0121 17:57:43.521085 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:43 crc kubenswrapper[4808]: E0121 17:57:43.521116 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:43 crc kubenswrapper[4808]: E0121 17:57:43.521433 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:43 crc kubenswrapper[4808]: E0121 17:57:43.521611 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:43 crc kubenswrapper[4808]: E0121 17:57:43.521694 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:45 crc kubenswrapper[4808]: I0121 17:57:45.520070 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:45 crc kubenswrapper[4808]: I0121 17:57:45.520150 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:45 crc kubenswrapper[4808]: I0121 17:57:45.520114 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:45 crc kubenswrapper[4808]: I0121 17:57:45.520107 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:45 crc kubenswrapper[4808]: E0121 17:57:45.522990 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:45 crc kubenswrapper[4808]: E0121 17:57:45.523342 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:45 crc kubenswrapper[4808]: E0121 17:57:45.523641 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:45 crc kubenswrapper[4808]: E0121 17:57:45.523923 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:47 crc kubenswrapper[4808]: I0121 17:57:47.400094 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs\") pod \"network-metrics-daemon-gm2t2\" (UID: \"8d39c565-f404-451e-896f-ecc5af76a000\") " pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:47 crc kubenswrapper[4808]: E0121 17:57:47.400369 4808 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 17:57:47 crc kubenswrapper[4808]: E0121 17:57:47.400453 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs podName:8d39c565-f404-451e-896f-ecc5af76a000 nodeName:}" failed. No retries permitted until 2026-01-21 17:58:51.400431744 +0000 UTC m=+166.130942659 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs") pod "network-metrics-daemon-gm2t2" (UID: "8d39c565-f404-451e-896f-ecc5af76a000") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 17:57:47 crc kubenswrapper[4808]: I0121 17:57:47.520082 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:47 crc kubenswrapper[4808]: I0121 17:57:47.520142 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:47 crc kubenswrapper[4808]: I0121 17:57:47.520075 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:47 crc kubenswrapper[4808]: E0121 17:57:47.520358 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:47 crc kubenswrapper[4808]: I0121 17:57:47.520489 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:47 crc kubenswrapper[4808]: E0121 17:57:47.520638 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:47 crc kubenswrapper[4808]: E0121 17:57:47.520733 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:47 crc kubenswrapper[4808]: E0121 17:57:47.521038 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:49 crc kubenswrapper[4808]: I0121 17:57:49.520416 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:49 crc kubenswrapper[4808]: I0121 17:57:49.520544 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:49 crc kubenswrapper[4808]: I0121 17:57:49.521523 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:49 crc kubenswrapper[4808]: E0121 17:57:49.521699 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:49 crc kubenswrapper[4808]: I0121 17:57:49.521748 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:49 crc kubenswrapper[4808]: E0121 17:57:49.521927 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:49 crc kubenswrapper[4808]: E0121 17:57:49.522044 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:49 crc kubenswrapper[4808]: E0121 17:57:49.522293 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:51 crc kubenswrapper[4808]: I0121 17:57:51.520055 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:51 crc kubenswrapper[4808]: I0121 17:57:51.520157 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:51 crc kubenswrapper[4808]: I0121 17:57:51.520187 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:51 crc kubenswrapper[4808]: E0121 17:57:51.520354 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:51 crc kubenswrapper[4808]: I0121 17:57:51.520417 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:51 crc kubenswrapper[4808]: E0121 17:57:51.520473 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:51 crc kubenswrapper[4808]: E0121 17:57:51.520633 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:51 crc kubenswrapper[4808]: E0121 17:57:51.520813 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:53 crc kubenswrapper[4808]: I0121 17:57:53.520502 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:53 crc kubenswrapper[4808]: I0121 17:57:53.520597 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:53 crc kubenswrapper[4808]: E0121 17:57:53.520714 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:53 crc kubenswrapper[4808]: I0121 17:57:53.520804 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:53 crc kubenswrapper[4808]: E0121 17:57:53.520882 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:53 crc kubenswrapper[4808]: I0121 17:57:53.520530 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:53 crc kubenswrapper[4808]: E0121 17:57:53.521040 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:53 crc kubenswrapper[4808]: E0121 17:57:53.521172 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:55 crc kubenswrapper[4808]: I0121 17:57:55.520398 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:55 crc kubenswrapper[4808]: I0121 17:57:55.521920 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:55 crc kubenswrapper[4808]: I0121 17:57:55.522041 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:55 crc kubenswrapper[4808]: E0121 17:57:55.522132 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:55 crc kubenswrapper[4808]: I0121 17:57:55.522361 4808 scope.go:117] "RemoveContainer" containerID="02b5440e1f9434c7998a819e1e62e4d10ad6887724bac037b63d65847c039f48" Jan 21 17:57:55 crc kubenswrapper[4808]: I0121 17:57:55.522389 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:55 crc kubenswrapper[4808]: E0121 17:57:55.522474 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:55 crc kubenswrapper[4808]: E0121 17:57:55.522589 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fc2fk_openshift-ovn-kubernetes(bb13a20f-f4ab-412c-8165-5b93edc79628)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" Jan 21 17:57:55 crc kubenswrapper[4808]: E0121 17:57:55.522844 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:55 crc kubenswrapper[4808]: E0121 17:57:55.522921 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:57 crc kubenswrapper[4808]: I0121 17:57:57.520399 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:57 crc kubenswrapper[4808]: I0121 17:57:57.520537 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:57 crc kubenswrapper[4808]: I0121 17:57:57.520399 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:57 crc kubenswrapper[4808]: I0121 17:57:57.520636 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:57 crc kubenswrapper[4808]: E0121 17:57:57.520628 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:57 crc kubenswrapper[4808]: E0121 17:57:57.520991 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:57:57 crc kubenswrapper[4808]: E0121 17:57:57.521141 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:57 crc kubenswrapper[4808]: E0121 17:57:57.520843 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:59 crc kubenswrapper[4808]: I0121 17:57:59.520748 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:57:59 crc kubenswrapper[4808]: I0121 17:57:59.520857 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:57:59 crc kubenswrapper[4808]: I0121 17:57:59.520891 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:57:59 crc kubenswrapper[4808]: E0121 17:57:59.521081 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:57:59 crc kubenswrapper[4808]: I0121 17:57:59.521516 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:57:59 crc kubenswrapper[4808]: E0121 17:57:59.521597 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:57:59 crc kubenswrapper[4808]: E0121 17:57:59.521725 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:57:59 crc kubenswrapper[4808]: E0121 17:57:59.521875 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:58:01 crc kubenswrapper[4808]: I0121 17:58:01.520228 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:58:01 crc kubenswrapper[4808]: I0121 17:58:01.520372 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:58:01 crc kubenswrapper[4808]: I0121 17:58:01.520388 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:58:01 crc kubenswrapper[4808]: I0121 17:58:01.520685 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:58:01 crc kubenswrapper[4808]: E0121 17:58:01.520683 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:58:01 crc kubenswrapper[4808]: E0121 17:58:01.520815 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:58:01 crc kubenswrapper[4808]: E0121 17:58:01.520978 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:58:01 crc kubenswrapper[4808]: E0121 17:58:01.521064 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:58:03 crc kubenswrapper[4808]: I0121 17:58:03.519766 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:58:03 crc kubenswrapper[4808]: I0121 17:58:03.519846 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:58:03 crc kubenswrapper[4808]: I0121 17:58:03.519916 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:58:03 crc kubenswrapper[4808]: E0121 17:58:03.519988 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:58:03 crc kubenswrapper[4808]: I0121 17:58:03.520029 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:58:03 crc kubenswrapper[4808]: E0121 17:58:03.520230 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:58:03 crc kubenswrapper[4808]: E0121 17:58:03.520485 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:58:03 crc kubenswrapper[4808]: E0121 17:58:03.520623 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:58:04 crc kubenswrapper[4808]: I0121 17:58:04.155804 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8fp9_538d1d44-6c67-450a-809c-d9008128ec9f/kube-multus/1.log" Jan 21 17:58:04 crc kubenswrapper[4808]: I0121 17:58:04.156457 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8fp9_538d1d44-6c67-450a-809c-d9008128ec9f/kube-multus/0.log" Jan 21 17:58:04 crc kubenswrapper[4808]: I0121 17:58:04.156529 4808 generic.go:334] "Generic (PLEG): container finished" podID="538d1d44-6c67-450a-809c-d9008128ec9f" containerID="f9074967ffe383fc367aaa834087e74f523d9563ecae59d79810bb389f05fbdc" exitCode=1 Jan 21 17:58:04 crc kubenswrapper[4808]: I0121 17:58:04.156582 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8fp9" event={"ID":"538d1d44-6c67-450a-809c-d9008128ec9f","Type":"ContainerDied","Data":"f9074967ffe383fc367aaa834087e74f523d9563ecae59d79810bb389f05fbdc"} Jan 21 17:58:04 crc kubenswrapper[4808]: I0121 17:58:04.156650 4808 scope.go:117] "RemoveContainer" containerID="33ae7632f3d0a28f819d32f4c459254bdb818a4f4d5b9b5681e324f5ffa04397" Jan 21 17:58:04 crc kubenswrapper[4808]: I0121 17:58:04.157323 4808 scope.go:117] "RemoveContainer" containerID="f9074967ffe383fc367aaa834087e74f523d9563ecae59d79810bb389f05fbdc" Jan 21 17:58:04 crc kubenswrapper[4808]: E0121 17:58:04.157503 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-z8fp9_openshift-multus(538d1d44-6c67-450a-809c-d9008128ec9f)\"" pod="openshift-multus/multus-z8fp9" podUID="538d1d44-6c67-450a-809c-d9008128ec9f" Jan 21 17:58:05 crc kubenswrapper[4808]: I0121 17:58:05.161703 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8fp9_538d1d44-6c67-450a-809c-d9008128ec9f/kube-multus/1.log" Jan 21 17:58:05 crc kubenswrapper[4808]: E0121 17:58:05.515074 4808 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Jan 21 17:58:05 crc kubenswrapper[4808]: I0121 17:58:05.519703 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:58:05 crc kubenswrapper[4808]: I0121 17:58:05.519781 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:58:05 crc kubenswrapper[4808]: E0121 17:58:05.521014 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:58:05 crc kubenswrapper[4808]: I0121 17:58:05.521051 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:58:05 crc kubenswrapper[4808]: I0121 17:58:05.521146 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:58:05 crc kubenswrapper[4808]: E0121 17:58:05.521219 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:58:05 crc kubenswrapper[4808]: E0121 17:58:05.521419 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:58:05 crc kubenswrapper[4808]: E0121 17:58:05.521458 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:58:05 crc kubenswrapper[4808]: E0121 17:58:05.615544 4808 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 21 17:58:07 crc kubenswrapper[4808]: I0121 17:58:07.520120 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:58:07 crc kubenswrapper[4808]: I0121 17:58:07.520144 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:58:07 crc kubenswrapper[4808]: I0121 17:58:07.520216 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:58:07 crc kubenswrapper[4808]: E0121 17:58:07.520385 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:58:07 crc kubenswrapper[4808]: I0121 17:58:07.520452 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:58:07 crc kubenswrapper[4808]: E0121 17:58:07.520668 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:58:07 crc kubenswrapper[4808]: E0121 17:58:07.520802 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:58:07 crc kubenswrapper[4808]: E0121 17:58:07.520886 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:58:08 crc kubenswrapper[4808]: I0121 17:58:08.521565 4808 scope.go:117] "RemoveContainer" containerID="02b5440e1f9434c7998a819e1e62e4d10ad6887724bac037b63d65847c039f48" Jan 21 17:58:09 crc kubenswrapper[4808]: I0121 17:58:09.182118 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovnkube-controller/3.log" Jan 21 17:58:09 crc kubenswrapper[4808]: I0121 17:58:09.184613 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerStarted","Data":"ac3a70f942488a03cabf1cce2d89342bfda3f785a417e6ec0246ff1a1f715c71"} Jan 21 17:58:09 crc kubenswrapper[4808]: I0121 17:58:09.185009 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:58:09 crc kubenswrapper[4808]: I0121 17:58:09.215902 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" podStartSLOduration=100.215865081 podStartE2EDuration="1m40.215865081s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:09.21409082 +0000 UTC m=+123.944601725" watchObservedRunningTime="2026-01-21 17:58:09.215865081 +0000 UTC m=+123.946375996" Jan 21 17:58:09 crc kubenswrapper[4808]: I0121 17:58:09.520460 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:58:09 crc kubenswrapper[4808]: I0121 17:58:09.520502 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:58:09 crc kubenswrapper[4808]: I0121 17:58:09.520545 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:58:09 crc kubenswrapper[4808]: I0121 17:58:09.520460 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:58:09 crc kubenswrapper[4808]: E0121 17:58:09.520597 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:58:09 crc kubenswrapper[4808]: E0121 17:58:09.520655 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:58:09 crc kubenswrapper[4808]: E0121 17:58:09.520794 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:58:09 crc kubenswrapper[4808]: E0121 17:58:09.520910 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:58:09 crc kubenswrapper[4808]: I0121 17:58:09.532046 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-gm2t2"] Jan 21 17:58:10 crc kubenswrapper[4808]: I0121 17:58:10.187126 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:58:10 crc kubenswrapper[4808]: E0121 17:58:10.187259 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:58:10 crc kubenswrapper[4808]: E0121 17:58:10.616427 4808 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 21 17:58:11 crc kubenswrapper[4808]: I0121 17:58:11.520615 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:58:11 crc kubenswrapper[4808]: I0121 17:58:11.520730 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:58:11 crc kubenswrapper[4808]: E0121 17:58:11.520804 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:58:11 crc kubenswrapper[4808]: E0121 17:58:11.520936 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:58:11 crc kubenswrapper[4808]: I0121 17:58:11.521034 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:58:11 crc kubenswrapper[4808]: E0121 17:58:11.521136 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:58:11 crc kubenswrapper[4808]: I0121 17:58:11.521157 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:58:11 crc kubenswrapper[4808]: E0121 17:58:11.521229 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:58:13 crc kubenswrapper[4808]: I0121 17:58:13.520222 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:58:13 crc kubenswrapper[4808]: I0121 17:58:13.520362 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:58:13 crc kubenswrapper[4808]: I0121 17:58:13.520401 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:58:13 crc kubenswrapper[4808]: E0121 17:58:13.521779 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:58:13 crc kubenswrapper[4808]: E0121 17:58:13.522075 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:58:13 crc kubenswrapper[4808]: E0121 17:58:13.522196 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:58:13 crc kubenswrapper[4808]: I0121 17:58:13.522840 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:58:13 crc kubenswrapper[4808]: E0121 17:58:13.523125 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:58:15 crc kubenswrapper[4808]: I0121 17:58:15.520616 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:58:15 crc kubenswrapper[4808]: E0121 17:58:15.521660 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:58:15 crc kubenswrapper[4808]: I0121 17:58:15.521727 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:58:15 crc kubenswrapper[4808]: I0121 17:58:15.521753 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:58:15 crc kubenswrapper[4808]: E0121 17:58:15.521945 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:58:15 crc kubenswrapper[4808]: E0121 17:58:15.522025 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:58:15 crc kubenswrapper[4808]: I0121 17:58:15.522066 4808 scope.go:117] "RemoveContainer" containerID="f9074967ffe383fc367aaa834087e74f523d9563ecae59d79810bb389f05fbdc" Jan 21 17:58:15 crc kubenswrapper[4808]: I0121 17:58:15.522289 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:58:15 crc kubenswrapper[4808]: E0121 17:58:15.522469 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:58:15 crc kubenswrapper[4808]: E0121 17:58:15.617163 4808 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 21 17:58:16 crc kubenswrapper[4808]: I0121 17:58:16.210177 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8fp9_538d1d44-6c67-450a-809c-d9008128ec9f/kube-multus/1.log" Jan 21 17:58:16 crc kubenswrapper[4808]: I0121 17:58:16.210269 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8fp9" event={"ID":"538d1d44-6c67-450a-809c-d9008128ec9f","Type":"ContainerStarted","Data":"e9f2431a8d4f410cded26b98a893f5432e81290e6947a1807e97cb72619b8c01"} Jan 21 17:58:17 crc kubenswrapper[4808]: I0121 17:58:17.519789 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:58:17 crc kubenswrapper[4808]: I0121 17:58:17.519866 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:58:17 crc kubenswrapper[4808]: I0121 17:58:17.519865 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:58:17 crc kubenswrapper[4808]: I0121 17:58:17.519935 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:58:17 crc kubenswrapper[4808]: E0121 17:58:17.520119 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:58:17 crc kubenswrapper[4808]: E0121 17:58:17.520510 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:58:17 crc kubenswrapper[4808]: E0121 17:58:17.520628 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:58:17 crc kubenswrapper[4808]: E0121 17:58:17.520802 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:58:19 crc kubenswrapper[4808]: I0121 17:58:19.520292 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:58:19 crc kubenswrapper[4808]: I0121 17:58:19.520416 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:58:19 crc kubenswrapper[4808]: E0121 17:58:19.520552 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 17:58:19 crc kubenswrapper[4808]: I0121 17:58:19.520574 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:58:19 crc kubenswrapper[4808]: I0121 17:58:19.520700 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:58:19 crc kubenswrapper[4808]: E0121 17:58:19.520861 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 17:58:19 crc kubenswrapper[4808]: E0121 17:58:19.521007 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 17:58:19 crc kubenswrapper[4808]: E0121 17:58:19.521137 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gm2t2" podUID="8d39c565-f404-451e-896f-ecc5af76a000" Jan 21 17:58:21 crc kubenswrapper[4808]: I0121 17:58:21.520216 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:58:21 crc kubenswrapper[4808]: I0121 17:58:21.520230 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:58:21 crc kubenswrapper[4808]: I0121 17:58:21.520462 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:58:21 crc kubenswrapper[4808]: I0121 17:58:21.520440 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:58:21 crc kubenswrapper[4808]: I0121 17:58:21.523791 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 21 17:58:21 crc kubenswrapper[4808]: I0121 17:58:21.524208 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 21 17:58:21 crc kubenswrapper[4808]: I0121 17:58:21.524502 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 21 17:58:21 crc kubenswrapper[4808]: I0121 17:58:21.524778 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 21 17:58:21 crc kubenswrapper[4808]: I0121 17:58:21.524833 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 21 17:58:21 crc kubenswrapper[4808]: I0121 17:58:21.524898 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.841680 4808 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.899113 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2q8lz"] Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.900386 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.900712 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx"] Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.901386 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.910345 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.910596 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.916793 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.918622 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.918909 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.919042 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.919181 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.919388 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.919499 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.919616 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.919709 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.919902 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.920287 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.931150 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.931509 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.931619 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.932163 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.932418 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.932605 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-etcd-client\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.932693 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-encryption-config\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.932767 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-serving-cert\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.932841 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-serving-cert\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.932895 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-audit-policies\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.932976 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-node-pullsecrets\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.933028 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-config\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.933156 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-etcd-client\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.933219 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-encryption-config\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.933300 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrwrm\" (UniqueName: \"kubernetes.io/projected/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-kube-api-access-zrwrm\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.933347 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.933363 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-audit-dir\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.933419 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.933473 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srr7v\" (UniqueName: \"kubernetes.io/projected/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-kube-api-access-srr7v\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.933522 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-audit\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.933581 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-etcd-serving-ca\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.933625 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-image-import-ca\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.933674 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.933759 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.933813 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-audit-dir\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.935468 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8"] Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.936373 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.937911 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.943105 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-ps7kv"] Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.950143 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.951216 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.950417 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ps7kv" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.950327 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b"] Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.951428 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.950821 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.951007 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.951572 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.952369 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4g7bd"] Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.952568 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.952771 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8h66x"] Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.953056 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hr5cr"] Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.953447 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-4g7bd" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.953555 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hr5cr" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.954015 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.961198 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.961821 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.962178 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.962184 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.962411 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.962335 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.964070 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.964988 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.966637 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.966974 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.967261 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.967415 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.967563 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.967560 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.967693 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.967712 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.967757 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.967806 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.967867 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.967899 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.967932 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.967954 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.967988 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.968044 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.968128 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.968166 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.968336 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.980362 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bz45s"] Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.980426 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.981040 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-fc6zr"] Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.981592 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-fc6zr" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.981611 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.985901 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-ftpvl"] Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.987668 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-h6gm7"] Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.987948 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-ftpvl" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.988853 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jcnqg"] Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.989027 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.989767 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-jcnqg" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.992145 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lq8cg"] Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.993928 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 21 17:58:29 crc kubenswrapper[4808]: I0121 17:58:29.994203 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.004878 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.007294 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cn7qm"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.010570 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2q8lz"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.012309 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-628ks"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.012466 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.013408 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.013530 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.012512 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lq8cg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.013319 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.014411 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cn7qm" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.022461 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.023307 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.035992 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.037047 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.037160 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.037225 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.037336 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.037434 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.037524 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.037636 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.037730 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.037905 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.037234 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.038090 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.038115 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.038302 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.038551 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jl78z"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.038811 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.039730 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jl78z" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.040709 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.041433 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.041637 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.042215 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.042418 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.042568 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043140 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/05688b08-2c9f-4c6c-967a-ff64de7837c0-console-config\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043191 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-node-pullsecrets\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043217 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-config\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043268 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043296 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-serving-cert\") pod \"route-controller-manager-6576b87f9c-wn2m8\" (UID: \"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043326 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05d34d86-bf40-4b28-94b9-8292fe685637-serving-cert\") pod \"authentication-operator-69f744f599-jcnqg\" (UID: \"05d34d86-bf40-4b28-94b9-8292fe685637\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jcnqg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043350 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-589nd\" (UniqueName: \"kubernetes.io/projected/3919133e-aaba-4dc1-b4cb-6b09f6440221-kube-api-access-589nd\") pod \"controller-manager-879f6c89f-8h66x\" (UID: \"3919133e-aaba-4dc1-b4cb-6b09f6440221\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043373 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/05688b08-2c9f-4c6c-967a-ff64de7837c0-console-oauth-config\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043397 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-etcd-client\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043402 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-node-pullsecrets\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043417 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-encryption-config\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043486 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrwrm\" (UniqueName: \"kubernetes.io/projected/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-kube-api-access-zrwrm\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043540 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8cb6\" (UniqueName: \"kubernetes.io/projected/05688b08-2c9f-4c6c-967a-ff64de7837c0-kube-api-access-c8cb6\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043576 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvdcv\" (UniqueName: \"kubernetes.io/projected/a189dfc0-4fe2-4290-b881-f238a055ab5a-kube-api-access-lvdcv\") pod \"console-operator-58897d9998-fc6zr\" (UID: \"a189dfc0-4fe2-4290-b881-f238a055ab5a\") " pod="openshift-console-operator/console-operator-58897d9998-fc6zr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043601 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ab1aa451-ae2a-48ec-ae13-0689bad3c894-auth-proxy-config\") pod \"machine-approver-56656f9798-ps7kv\" (UID: \"ab1aa451-ae2a-48ec-ae13-0689bad3c894\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ps7kv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043623 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7zt6\" (UniqueName: \"kubernetes.io/projected/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-kube-api-access-x7zt6\") pod \"route-controller-manager-6576b87f9c-wn2m8\" (UID: \"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043649 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26022c12-cb37-40ed-9ecb-aef5f82d2bed-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-hr5cr\" (UID: \"26022c12-cb37-40ed-9ecb-aef5f82d2bed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hr5cr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043672 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a189dfc0-4fe2-4290-b881-f238a055ab5a-trusted-ca\") pod \"console-operator-58897d9998-fc6zr\" (UID: \"a189dfc0-4fe2-4290-b881-f238a055ab5a\") " pod="openshift-console-operator/console-operator-58897d9998-fc6zr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043702 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-audit-dir\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043727 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05d34d86-bf40-4b28-94b9-8292fe685637-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jcnqg\" (UID: \"05d34d86-bf40-4b28-94b9-8292fe685637\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jcnqg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043749 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a189dfc0-4fe2-4290-b881-f238a055ab5a-serving-cert\") pod \"console-operator-58897d9998-fc6zr\" (UID: \"a189dfc0-4fe2-4290-b881-f238a055ab5a\") " pod="openshift-console-operator/console-operator-58897d9998-fc6zr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043774 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043800 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043829 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a189dfc0-4fe2-4290-b881-f238a055ab5a-config\") pod \"console-operator-58897d9998-fc6zr\" (UID: \"a189dfc0-4fe2-4290-b881-f238a055ab5a\") " pod="openshift-console-operator/console-operator-58897d9998-fc6zr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043860 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srr7v\" (UniqueName: \"kubernetes.io/projected/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-kube-api-access-srr7v\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043885 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043911 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c1e9bbd-24f9-4ef3-8844-6ddb9861c509-serving-cert\") pod \"openshift-config-operator-7777fb866f-p9d2b\" (UID: \"1c1e9bbd-24f9-4ef3-8844-6ddb9861c509\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043939 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxrs7\" (UniqueName: \"kubernetes.io/projected/1c1e9bbd-24f9-4ef3-8844-6ddb9861c509-kube-api-access-jxrs7\") pod \"openshift-config-operator-7777fb866f-p9d2b\" (UID: \"1c1e9bbd-24f9-4ef3-8844-6ddb9861c509\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043967 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-audit\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.043993 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab1aa451-ae2a-48ec-ae13-0689bad3c894-config\") pod \"machine-approver-56656f9798-ps7kv\" (UID: \"ab1aa451-ae2a-48ec-ae13-0689bad3c894\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ps7kv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044018 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54slc\" (UniqueName: \"kubernetes.io/projected/ab1aa451-ae2a-48ec-ae13-0689bad3c894-kube-api-access-54slc\") pod \"machine-approver-56656f9798-ps7kv\" (UID: \"ab1aa451-ae2a-48ec-ae13-0689bad3c894\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ps7kv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044046 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3919133e-aaba-4dc1-b4cb-6b09f6440221-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8h66x\" (UID: \"3919133e-aaba-4dc1-b4cb-6b09f6440221\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044071 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b91a9d5-e508-4624-a43c-770a9dcdbc1d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lq8cg\" (UID: \"5b91a9d5-e508-4624-a43c-770a9dcdbc1d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lq8cg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044098 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044120 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044146 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044173 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/05688b08-2c9f-4c6c-967a-ff64de7837c0-console-serving-cert\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044202 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-etcd-serving-ca\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044225 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-image-import-ca\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044272 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044302 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n95rc\" (UniqueName: \"kubernetes.io/projected/5b91a9d5-e508-4624-a43c-770a9dcdbc1d-kube-api-access-n95rc\") pod \"cluster-image-registry-operator-dc59b4c8b-lq8cg\" (UID: \"5b91a9d5-e508-4624-a43c-770a9dcdbc1d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lq8cg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044330 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4c87e4f3-6255-4fe3-a52c-726a1acf7ba8-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4g7bd\" (UID: \"4c87e4f3-6255-4fe3-a52c-726a1acf7ba8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g7bd" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044353 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5c7z\" (UniqueName: \"kubernetes.io/projected/4c87e4f3-6255-4fe3-a52c-726a1acf7ba8-kube-api-access-s5c7z\") pod \"machine-api-operator-5694c8668f-4g7bd\" (UID: \"4c87e4f3-6255-4fe3-a52c-726a1acf7ba8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g7bd" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044398 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3919133e-aaba-4dc1-b4cb-6b09f6440221-client-ca\") pod \"controller-manager-879f6c89f-8h66x\" (UID: \"3919133e-aaba-4dc1-b4cb-6b09f6440221\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044422 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-audit-policies\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044463 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044489 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-audit-dir\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044514 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3919133e-aaba-4dc1-b4cb-6b09f6440221-config\") pod \"controller-manager-879f6c89f-8h66x\" (UID: \"3919133e-aaba-4dc1-b4cb-6b09f6440221\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044540 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ab1aa451-ae2a-48ec-ae13-0689bad3c894-machine-approver-tls\") pod \"machine-approver-56656f9798-ps7kv\" (UID: \"ab1aa451-ae2a-48ec-ae13-0689bad3c894\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ps7kv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044563 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044591 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044621 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05d34d86-bf40-4b28-94b9-8292fe685637-service-ca-bundle\") pod \"authentication-operator-69f744f599-jcnqg\" (UID: \"05d34d86-bf40-4b28-94b9-8292fe685637\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jcnqg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044645 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3919133e-aaba-4dc1-b4cb-6b09f6440221-serving-cert\") pod \"controller-manager-879f6c89f-8h66x\" (UID: \"3919133e-aaba-4dc1-b4cb-6b09f6440221\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044686 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4c87e4f3-6255-4fe3-a52c-726a1acf7ba8-images\") pod \"machine-api-operator-5694c8668f-4g7bd\" (UID: \"4c87e4f3-6255-4fe3-a52c-726a1acf7ba8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g7bd" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044714 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05d34d86-bf40-4b28-94b9-8292fe685637-config\") pod \"authentication-operator-69f744f599-jcnqg\" (UID: \"05d34d86-bf40-4b28-94b9-8292fe685637\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jcnqg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044753 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vzpx\" (UniqueName: \"kubernetes.io/projected/05d34d86-bf40-4b28-94b9-8292fe685637-kube-api-access-7vzpx\") pod \"authentication-operator-69f744f599-jcnqg\" (UID: \"05d34d86-bf40-4b28-94b9-8292fe685637\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jcnqg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044779 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skpzt\" (UniqueName: \"kubernetes.io/projected/66c5c3c8-4515-401a-bfe9-5755068bc771-kube-api-access-skpzt\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044807 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/5b91a9d5-e508-4624-a43c-770a9dcdbc1d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lq8cg\" (UID: \"5b91a9d5-e508-4624-a43c-770a9dcdbc1d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lq8cg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044832 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-config\") pod \"route-controller-manager-6576b87f9c-wn2m8\" (UID: \"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044858 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044880 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/05688b08-2c9f-4c6c-967a-ff64de7837c0-service-ca\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044904 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/05688b08-2c9f-4c6c-967a-ff64de7837c0-oauth-serving-cert\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044941 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-etcd-client\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044965 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26022c12-cb37-40ed-9ecb-aef5f82d2bed-config\") pod \"openshift-apiserver-operator-796bbdcf4f-hr5cr\" (UID: \"26022c12-cb37-40ed-9ecb-aef5f82d2bed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hr5cr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.044989 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5b91a9d5-e508-4624-a43c-770a9dcdbc1d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lq8cg\" (UID: \"5b91a9d5-e508-4624-a43c-770a9dcdbc1d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lq8cg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.045017 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/1c1e9bbd-24f9-4ef3-8844-6ddb9861c509-available-featuregates\") pod \"openshift-config-operator-7777fb866f-p9d2b\" (UID: \"1c1e9bbd-24f9-4ef3-8844-6ddb9861c509\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.045069 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-encryption-config\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.045098 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66c5c3c8-4515-401a-bfe9-5755068bc771-audit-dir\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.045128 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-serving-cert\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.045170 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-serving-cert\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.045205 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlc27\" (UniqueName: \"kubernetes.io/projected/26022c12-cb37-40ed-9ecb-aef5f82d2bed-kube-api-access-dlc27\") pod \"openshift-apiserver-operator-796bbdcf4f-hr5cr\" (UID: \"26022c12-cb37-40ed-9ecb-aef5f82d2bed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hr5cr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.045254 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.045285 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c87e4f3-6255-4fe3-a52c-726a1acf7ba8-config\") pod \"machine-api-operator-5694c8668f-4g7bd\" (UID: \"4c87e4f3-6255-4fe3-a52c-726a1acf7ba8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g7bd" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.045318 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-audit-policies\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.045346 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d992\" (UniqueName: \"kubernetes.io/projected/196e1220-375f-4a05-90e0-ce571f606de4-kube-api-access-5d992\") pod \"downloads-7954f5f757-ftpvl\" (UID: \"196e1220-375f-4a05-90e0-ce571f606de4\") " pod="openshift-console/downloads-7954f5f757-ftpvl" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.045361 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.045373 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.045607 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-client-ca\") pod \"route-controller-manager-6576b87f9c-wn2m8\" (UID: \"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.045642 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05688b08-2c9f-4c6c-967a-ff64de7837c0-trusted-ca-bundle\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.046451 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-audit\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.046820 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-etcd-serving-ca\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.048302 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.050159 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.050318 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.050894 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.051332 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.051562 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.053773 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-image-import-ca\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.054193 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.054548 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.055161 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-config\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.056311 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-etcd-client\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.056395 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-audit-dir\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.057389 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-audit-dir\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.058005 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-audit-policies\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.058123 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-slth9"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.059058 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-slth9" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.061281 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.064294 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.064530 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.065322 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-encryption-config\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.065768 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-serving-cert\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.065849 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8jqv"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.066033 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.066197 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-serving-cert\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.066524 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8jqv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.069074 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-encryption-config\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.069227 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.069923 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.071177 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.071646 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.072572 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-etcd-client\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.074514 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.075717 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.076212 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.078708 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.078861 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xcn7d"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.079607 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.079711 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xcn7d" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.092748 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.094312 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.095739 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-q95v9"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.096827 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-n56cl"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.098398 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.100612 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.101353 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.106637 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q95v9" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.106768 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.109865 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wj246"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.110918 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wj246" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.112450 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.118921 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-blj6k"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.124430 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-blj6k" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.126386 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-jxgs8"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.127740 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z77hq"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.127885 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-jxgs8" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.128627 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.129267 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-z77hq" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.130844 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.130874 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-xfblx"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.131836 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.132068 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.132179 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.132731 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.133087 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jmp6c"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.134056 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q8wfq"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.134672 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jmp6c" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.136565 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.137014 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.137521 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4gg7j"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.137879 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-s5khx"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.137966 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q8wfq" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.138003 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.137971 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4gg7j" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.141767 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fn8s9"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.142004 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-s5khx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.142390 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-d2rkk"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.142464 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.142587 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fn8s9" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.142936 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.142962 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pvk7n"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.143254 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d2rkk" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.143469 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4g7bd"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.143689 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.144359 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-h6gm7"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.145421 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.146320 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hr5cr"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.147220 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lpz9\" (UniqueName: \"kubernetes.io/projected/871099ce-cd86-498e-b227-ec40b6c29e47-kube-api-access-2lpz9\") pod \"openshift-controller-manager-operator-756b6f6bc6-q8jqv\" (UID: \"871099ce-cd86-498e-b227-ec40b6c29e47\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8jqv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.147454 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/05688b08-2c9f-4c6c-967a-ff64de7837c0-console-config\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.147583 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-glhp5"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.147594 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.147800 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z98n2\" (UniqueName: \"kubernetes.io/projected/56d0bac1-fb03-49ab-8e45-1cb28643dc04-kube-api-access-z98n2\") pod \"etcd-operator-b45778765-628ks\" (UID: \"56d0bac1-fb03-49ab-8e45-1cb28643dc04\") " pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.147917 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-serving-cert\") pod \"route-controller-manager-6576b87f9c-wn2m8\" (UID: \"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.148027 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05d34d86-bf40-4b28-94b9-8292fe685637-serving-cert\") pod \"authentication-operator-69f744f599-jcnqg\" (UID: \"05d34d86-bf40-4b28-94b9-8292fe685637\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jcnqg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.148132 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-589nd\" (UniqueName: \"kubernetes.io/projected/3919133e-aaba-4dc1-b4cb-6b09f6440221-kube-api-access-589nd\") pod \"controller-manager-879f6c89f-8h66x\" (UID: \"3919133e-aaba-4dc1-b4cb-6b09f6440221\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.148265 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/05688b08-2c9f-4c6c-967a-ff64de7837c0-console-oauth-config\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.148468 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afb48788-1b96-4ba0-9a9d-98e613451c6f-config\") pod \"kube-apiserver-operator-766d6c64bb-blj6k\" (UID: \"afb48788-1b96-4ba0-9a9d-98e613451c6f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-blj6k" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.148593 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8cb6\" (UniqueName: \"kubernetes.io/projected/05688b08-2c9f-4c6c-967a-ff64de7837c0-kube-api-access-c8cb6\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.148346 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/05688b08-2c9f-4c6c-967a-ff64de7837c0-console-config\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.148901 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvdcv\" (UniqueName: \"kubernetes.io/projected/a189dfc0-4fe2-4290-b881-f238a055ab5a-kube-api-access-lvdcv\") pod \"console-operator-58897d9998-fc6zr\" (UID: \"a189dfc0-4fe2-4290-b881-f238a055ab5a\") " pod="openshift-console-operator/console-operator-58897d9998-fc6zr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.149170 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ab1aa451-ae2a-48ec-ae13-0689bad3c894-auth-proxy-config\") pod \"machine-approver-56656f9798-ps7kv\" (UID: \"ab1aa451-ae2a-48ec-ae13-0689bad3c894\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ps7kv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.149462 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7zt6\" (UniqueName: \"kubernetes.io/projected/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-kube-api-access-x7zt6\") pod \"route-controller-manager-6576b87f9c-wn2m8\" (UID: \"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.149595 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26022c12-cb37-40ed-9ecb-aef5f82d2bed-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-hr5cr\" (UID: \"26022c12-cb37-40ed-9ecb-aef5f82d2bed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hr5cr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.149697 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a189dfc0-4fe2-4290-b881-f238a055ab5a-trusted-ca\") pod \"console-operator-58897d9998-fc6zr\" (UID: \"a189dfc0-4fe2-4290-b881-f238a055ab5a\") " pod="openshift-console-operator/console-operator-58897d9998-fc6zr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.149790 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05d34d86-bf40-4b28-94b9-8292fe685637-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jcnqg\" (UID: \"05d34d86-bf40-4b28-94b9-8292fe685637\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jcnqg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.149906 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a189dfc0-4fe2-4290-b881-f238a055ab5a-serving-cert\") pod \"console-operator-58897d9998-fc6zr\" (UID: \"a189dfc0-4fe2-4290-b881-f238a055ab5a\") " pod="openshift-console-operator/console-operator-58897d9998-fc6zr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.150060 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.150157 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.150460 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a189dfc0-4fe2-4290-b881-f238a055ab5a-config\") pod \"console-operator-58897d9998-fc6zr\" (UID: \"a189dfc0-4fe2-4290-b881-f238a055ab5a\") " pod="openshift-console-operator/console-operator-58897d9998-fc6zr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.150569 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c1e9bbd-24f9-4ef3-8844-6ddb9861c509-serving-cert\") pod \"openshift-config-operator-7777fb866f-p9d2b\" (UID: \"1c1e9bbd-24f9-4ef3-8844-6ddb9861c509\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.150676 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/871099ce-cd86-498e-b227-ec40b6c29e47-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-q8jqv\" (UID: \"871099ce-cd86-498e-b227-ec40b6c29e47\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8jqv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.150120 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ab1aa451-ae2a-48ec-ae13-0689bad3c894-auth-proxy-config\") pod \"machine-approver-56656f9798-ps7kv\" (UID: \"ab1aa451-ae2a-48ec-ae13-0689bad3c894\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ps7kv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.150852 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc7s6\" (UniqueName: \"kubernetes.io/projected/11782075-8384-415d-a1b9-6c972f8f6f66-kube-api-access-bc7s6\") pod \"control-plane-machine-set-operator-78cbb6b69f-jl78z\" (UID: \"11782075-8384-415d-a1b9-6c972f8f6f66\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jl78z" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.150971 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxrs7\" (UniqueName: \"kubernetes.io/projected/1c1e9bbd-24f9-4ef3-8844-6ddb9861c509-kube-api-access-jxrs7\") pod \"openshift-config-operator-7777fb866f-p9d2b\" (UID: \"1c1e9bbd-24f9-4ef3-8844-6ddb9861c509\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.151129 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05d34d86-bf40-4b28-94b9-8292fe685637-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jcnqg\" (UID: \"05d34d86-bf40-4b28-94b9-8292fe685637\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jcnqg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.151141 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/871099ce-cd86-498e-b227-ec40b6c29e47-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-q8jqv\" (UID: \"871099ce-cd86-498e-b227-ec40b6c29e47\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8jqv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.150870 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-glhp5" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.151207 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab1aa451-ae2a-48ec-ae13-0689bad3c894-config\") pod \"machine-approver-56656f9798-ps7kv\" (UID: \"ab1aa451-ae2a-48ec-ae13-0689bad3c894\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ps7kv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.150720 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lq8cg"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.151303 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54slc\" (UniqueName: \"kubernetes.io/projected/ab1aa451-ae2a-48ec-ae13-0689bad3c894-kube-api-access-54slc\") pod \"machine-approver-56656f9798-ps7kv\" (UID: \"ab1aa451-ae2a-48ec-ae13-0689bad3c894\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ps7kv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.151854 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/56d0bac1-fb03-49ab-8e45-1cb28643dc04-etcd-service-ca\") pod \"etcd-operator-b45778765-628ks\" (UID: \"56d0bac1-fb03-49ab-8e45-1cb28643dc04\") " pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.151953 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b91a9d5-e508-4624-a43c-770a9dcdbc1d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lq8cg\" (UID: \"5b91a9d5-e508-4624-a43c-770a9dcdbc1d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lq8cg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152031 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3919133e-aaba-4dc1-b4cb-6b09f6440221-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8h66x\" (UID: \"3919133e-aaba-4dc1-b4cb-6b09f6440221\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152089 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzq7d\" (UniqueName: \"kubernetes.io/projected/dbc348aa-a6f2-4fee-8c96-83f98c8358b9-kube-api-access-mzq7d\") pod \"migrator-59844c95c7-q95v9\" (UID: \"dbc348aa-a6f2-4fee-8c96-83f98c8358b9\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q95v9" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152113 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/56d0bac1-fb03-49ab-8e45-1cb28643dc04-etcd-client\") pod \"etcd-operator-b45778765-628ks\" (UID: \"56d0bac1-fb03-49ab-8e45-1cb28643dc04\") " pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152175 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152204 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152229 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152286 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56d0bac1-fb03-49ab-8e45-1cb28643dc04-config\") pod \"etcd-operator-b45778765-628ks\" (UID: \"56d0bac1-fb03-49ab-8e45-1cb28643dc04\") " pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152316 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/05688b08-2c9f-4c6c-967a-ff64de7837c0-console-serving-cert\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152372 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n95rc\" (UniqueName: \"kubernetes.io/projected/5b91a9d5-e508-4624-a43c-770a9dcdbc1d-kube-api-access-n95rc\") pod \"cluster-image-registry-operator-dc59b4c8b-lq8cg\" (UID: \"5b91a9d5-e508-4624-a43c-770a9dcdbc1d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lq8cg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152432 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4c87e4f3-6255-4fe3-a52c-726a1acf7ba8-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4g7bd\" (UID: \"4c87e4f3-6255-4fe3-a52c-726a1acf7ba8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g7bd" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152444 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a189dfc0-4fe2-4290-b881-f238a055ab5a-config\") pod \"console-operator-58897d9998-fc6zr\" (UID: \"a189dfc0-4fe2-4290-b881-f238a055ab5a\") " pod="openshift-console-operator/console-operator-58897d9998-fc6zr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152460 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5c7z\" (UniqueName: \"kubernetes.io/projected/4c87e4f3-6255-4fe3-a52c-726a1acf7ba8-kube-api-access-s5c7z\") pod \"machine-api-operator-5694c8668f-4g7bd\" (UID: \"4c87e4f3-6255-4fe3-a52c-726a1acf7ba8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g7bd" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152517 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152525 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3919133e-aaba-4dc1-b4cb-6b09f6440221-client-ca\") pod \"controller-manager-879f6c89f-8h66x\" (UID: \"3919133e-aaba-4dc1-b4cb-6b09f6440221\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152602 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-audit-policies\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152645 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3919133e-aaba-4dc1-b4cb-6b09f6440221-config\") pod \"controller-manager-879f6c89f-8h66x\" (UID: \"3919133e-aaba-4dc1-b4cb-6b09f6440221\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152689 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ab1aa451-ae2a-48ec-ae13-0689bad3c894-machine-approver-tls\") pod \"machine-approver-56656f9798-ps7kv\" (UID: \"ab1aa451-ae2a-48ec-ae13-0689bad3c894\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ps7kv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152708 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152729 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152750 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4c87e4f3-6255-4fe3-a52c-726a1acf7ba8-images\") pod \"machine-api-operator-5694c8668f-4g7bd\" (UID: \"4c87e4f3-6255-4fe3-a52c-726a1acf7ba8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g7bd" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152772 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05d34d86-bf40-4b28-94b9-8292fe685637-service-ca-bundle\") pod \"authentication-operator-69f744f599-jcnqg\" (UID: \"05d34d86-bf40-4b28-94b9-8292fe685637\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jcnqg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152794 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3919133e-aaba-4dc1-b4cb-6b09f6440221-serving-cert\") pod \"controller-manager-879f6c89f-8h66x\" (UID: \"3919133e-aaba-4dc1-b4cb-6b09f6440221\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152817 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05d34d86-bf40-4b28-94b9-8292fe685637-config\") pod \"authentication-operator-69f744f599-jcnqg\" (UID: \"05d34d86-bf40-4b28-94b9-8292fe685637\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jcnqg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152842 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vzpx\" (UniqueName: \"kubernetes.io/projected/05d34d86-bf40-4b28-94b9-8292fe685637-kube-api-access-7vzpx\") pod \"authentication-operator-69f744f599-jcnqg\" (UID: \"05d34d86-bf40-4b28-94b9-8292fe685637\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jcnqg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152863 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skpzt\" (UniqueName: \"kubernetes.io/projected/66c5c3c8-4515-401a-bfe9-5755068bc771-kube-api-access-skpzt\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152866 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab1aa451-ae2a-48ec-ae13-0689bad3c894-config\") pod \"machine-approver-56656f9798-ps7kv\" (UID: \"ab1aa451-ae2a-48ec-ae13-0689bad3c894\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ps7kv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152901 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/5b91a9d5-e508-4624-a43c-770a9dcdbc1d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lq8cg\" (UID: \"5b91a9d5-e508-4624-a43c-770a9dcdbc1d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lq8cg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152922 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-config\") pod \"route-controller-manager-6576b87f9c-wn2m8\" (UID: \"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152944 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/05688b08-2c9f-4c6c-967a-ff64de7837c0-service-ca\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152963 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/05688b08-2c9f-4c6c-967a-ff64de7837c0-oauth-serving-cert\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.152990 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153036 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5b91a9d5-e508-4624-a43c-770a9dcdbc1d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lq8cg\" (UID: \"5b91a9d5-e508-4624-a43c-770a9dcdbc1d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lq8cg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153060 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/1c1e9bbd-24f9-4ef3-8844-6ddb9861c509-available-featuregates\") pod \"openshift-config-operator-7777fb866f-p9d2b\" (UID: \"1c1e9bbd-24f9-4ef3-8844-6ddb9861c509\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153085 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/afb48788-1b96-4ba0-9a9d-98e613451c6f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-blj6k\" (UID: \"afb48788-1b96-4ba0-9a9d-98e613451c6f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-blj6k" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153106 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d69wz\" (UniqueName: \"kubernetes.io/projected/a58048c2-af3b-4ea3-820e-622c72edd91f-kube-api-access-d69wz\") pod \"cluster-samples-operator-665b6dd947-cn7qm\" (UID: \"a58048c2-af3b-4ea3-820e-622c72edd91f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cn7qm" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153109 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3919133e-aaba-4dc1-b4cb-6b09f6440221-client-ca\") pod \"controller-manager-879f6c89f-8h66x\" (UID: \"3919133e-aaba-4dc1-b4cb-6b09f6440221\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153133 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26022c12-cb37-40ed-9ecb-aef5f82d2bed-config\") pod \"openshift-apiserver-operator-796bbdcf4f-hr5cr\" (UID: \"26022c12-cb37-40ed-9ecb-aef5f82d2bed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hr5cr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153159 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/11782075-8384-415d-a1b9-6c972f8f6f66-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jl78z\" (UID: \"11782075-8384-415d-a1b9-6c972f8f6f66\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jl78z" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153197 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66c5c3c8-4515-401a-bfe9-5755068bc771-audit-dir\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153220 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/56d0bac1-fb03-49ab-8e45-1cb28643dc04-etcd-ca\") pod \"etcd-operator-b45778765-628ks\" (UID: \"56d0bac1-fb03-49ab-8e45-1cb28643dc04\") " pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153287 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afb48788-1b96-4ba0-9a9d-98e613451c6f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-blj6k\" (UID: \"afb48788-1b96-4ba0-9a9d-98e613451c6f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-blj6k" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153309 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3919133e-aaba-4dc1-b4cb-6b09f6440221-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8h66x\" (UID: \"3919133e-aaba-4dc1-b4cb-6b09f6440221\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153310 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c87e4f3-6255-4fe3-a52c-726a1acf7ba8-config\") pod \"machine-api-operator-5694c8668f-4g7bd\" (UID: \"4c87e4f3-6255-4fe3-a52c-726a1acf7ba8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g7bd" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153391 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlc27\" (UniqueName: \"kubernetes.io/projected/26022c12-cb37-40ed-9ecb-aef5f82d2bed-kube-api-access-dlc27\") pod \"openshift-apiserver-operator-796bbdcf4f-hr5cr\" (UID: \"26022c12-cb37-40ed-9ecb-aef5f82d2bed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hr5cr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153411 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153437 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d992\" (UniqueName: \"kubernetes.io/projected/196e1220-375f-4a05-90e0-ce571f606de4-kube-api-access-5d992\") pod \"downloads-7954f5f757-ftpvl\" (UID: \"196e1220-375f-4a05-90e0-ce571f606de4\") " pod="openshift-console/downloads-7954f5f757-ftpvl" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153458 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153477 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-client-ca\") pod \"route-controller-manager-6576b87f9c-wn2m8\" (UID: \"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153496 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05688b08-2c9f-4c6c-967a-ff64de7837c0-trusted-ca-bundle\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153527 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a58048c2-af3b-4ea3-820e-622c72edd91f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-cn7qm\" (UID: \"a58048c2-af3b-4ea3-820e-622c72edd91f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cn7qm" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153529 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153545 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56d0bac1-fb03-49ab-8e45-1cb28643dc04-serving-cert\") pod \"etcd-operator-b45778765-628ks\" (UID: \"56d0bac1-fb03-49ab-8e45-1cb28643dc04\") " pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.153310 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b91a9d5-e508-4624-a43c-770a9dcdbc1d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lq8cg\" (UID: \"5b91a9d5-e508-4624-a43c-770a9dcdbc1d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lq8cg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.154530 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-config\") pod \"route-controller-manager-6576b87f9c-wn2m8\" (UID: \"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.155161 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c87e4f3-6255-4fe3-a52c-726a1acf7ba8-config\") pod \"machine-api-operator-5694c8668f-4g7bd\" (UID: \"4c87e4f3-6255-4fe3-a52c-726a1acf7ba8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g7bd" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.155223 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05d34d86-bf40-4b28-94b9-8292fe685637-service-ca-bundle\") pod \"authentication-operator-69f744f599-jcnqg\" (UID: \"05d34d86-bf40-4b28-94b9-8292fe685637\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jcnqg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.155548 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-client-ca\") pod \"route-controller-manager-6576b87f9c-wn2m8\" (UID: \"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.155792 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.156182 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/05688b08-2c9f-4c6c-967a-ff64de7837c0-oauth-serving-cert\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.156184 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.156214 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.156224 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.156410 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/05688b08-2c9f-4c6c-967a-ff64de7837c0-service-ca\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.156593 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/05688b08-2c9f-4c6c-967a-ff64de7837c0-console-oauth-config\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.156932 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05d34d86-bf40-4b28-94b9-8292fe685637-config\") pod \"authentication-operator-69f744f599-jcnqg\" (UID: \"05d34d86-bf40-4b28-94b9-8292fe685637\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jcnqg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.156976 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66c5c3c8-4515-401a-bfe9-5755068bc771-audit-dir\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.157002 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4c87e4f3-6255-4fe3-a52c-726a1acf7ba8-images\") pod \"machine-api-operator-5694c8668f-4g7bd\" (UID: \"4c87e4f3-6255-4fe3-a52c-726a1acf7ba8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g7bd" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.157036 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-serving-cert\") pod \"route-controller-manager-6576b87f9c-wn2m8\" (UID: \"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.157274 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26022c12-cb37-40ed-9ecb-aef5f82d2bed-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-hr5cr\" (UID: \"26022c12-cb37-40ed-9ecb-aef5f82d2bed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hr5cr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.157599 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26022c12-cb37-40ed-9ecb-aef5f82d2bed-config\") pod \"openshift-apiserver-operator-796bbdcf4f-hr5cr\" (UID: \"26022c12-cb37-40ed-9ecb-aef5f82d2bed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hr5cr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.157893 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/1c1e9bbd-24f9-4ef3-8844-6ddb9861c509-available-featuregates\") pod \"openshift-config-operator-7777fb866f-p9d2b\" (UID: \"1c1e9bbd-24f9-4ef3-8844-6ddb9861c509\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.158152 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.158197 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-audit-policies\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.158437 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a189dfc0-4fe2-4290-b881-f238a055ab5a-trusted-ca\") pod \"console-operator-58897d9998-fc6zr\" (UID: \"a189dfc0-4fe2-4290-b881-f238a055ab5a\") " pod="openshift-console-operator/console-operator-58897d9998-fc6zr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.159128 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05d34d86-bf40-4b28-94b9-8292fe685637-serving-cert\") pod \"authentication-operator-69f744f599-jcnqg\" (UID: \"05d34d86-bf40-4b28-94b9-8292fe685637\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jcnqg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.159374 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.159747 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05688b08-2c9f-4c6c-967a-ff64de7837c0-trusted-ca-bundle\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.159829 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.159895 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3919133e-aaba-4dc1-b4cb-6b09f6440221-config\") pod \"controller-manager-879f6c89f-8h66x\" (UID: \"3919133e-aaba-4dc1-b4cb-6b09f6440221\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.159968 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/05688b08-2c9f-4c6c-967a-ff64de7837c0-console-serving-cert\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.161414 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jl78z"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.161510 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.161816 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ab1aa451-ae2a-48ec-ae13-0689bad3c894-machine-approver-tls\") pod \"machine-approver-56656f9798-ps7kv\" (UID: \"ab1aa451-ae2a-48ec-ae13-0689bad3c894\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ps7kv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.161892 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.161913 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c1e9bbd-24f9-4ef3-8844-6ddb9861c509-serving-cert\") pod \"openshift-config-operator-7777fb866f-p9d2b\" (UID: \"1c1e9bbd-24f9-4ef3-8844-6ddb9861c509\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.161984 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a189dfc0-4fe2-4290-b881-f238a055ab5a-serving-cert\") pod \"console-operator-58897d9998-fc6zr\" (UID: \"a189dfc0-4fe2-4290-b881-f238a055ab5a\") " pod="openshift-console-operator/console-operator-58897d9998-fc6zr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.162069 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3919133e-aaba-4dc1-b4cb-6b09f6440221-serving-cert\") pod \"controller-manager-879f6c89f-8h66x\" (UID: \"3919133e-aaba-4dc1-b4cb-6b09f6440221\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.162393 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.163466 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4c87e4f3-6255-4fe3-a52c-726a1acf7ba8-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4g7bd\" (UID: \"4c87e4f3-6255-4fe3-a52c-726a1acf7ba8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g7bd" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.163534 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.165454 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-ftpvl"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.166493 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/5b91a9d5-e508-4624-a43c-770a9dcdbc1d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lq8cg\" (UID: \"5b91a9d5-e508-4624-a43c-770a9dcdbc1d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lq8cg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.167286 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8h66x"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.171585 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bz45s"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.172998 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cn7qm"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.175643 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z77hq"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.177042 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.178322 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-628ks"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.179370 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-blj6k"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.180393 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-slth9"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.181769 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jcnqg"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.182722 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4gg7j"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.184151 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xcn7d"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.188888 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-d2rkk"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.192059 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8jqv"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.194025 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-fc6zr"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.195526 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q8wfq"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.199464 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.200604 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-q95v9"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.200874 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrwrm\" (UniqueName: \"kubernetes.io/projected/9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d-kube-api-access-zrwrm\") pod \"apiserver-76f77b778f-2q8lz\" (UID: \"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d\") " pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.201922 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-zjcb6"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.202692 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.202715 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-zjcb6" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.203656 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-jxgs8"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.205283 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wj246"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.207020 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-n56cl"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.208979 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-zjcb6"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.209950 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pvk7n"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.211127 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jmp6c"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.212507 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-glhp5"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.214285 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.216408 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fn8s9"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.217853 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-s5khx"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.219123 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-2m5jx"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.219659 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2m5jx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.220455 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-79g2n"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.220912 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-79g2n" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.221880 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-79g2n"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.224417 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.243274 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.254410 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.254664 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z98n2\" (UniqueName: \"kubernetes.io/projected/56d0bac1-fb03-49ab-8e45-1cb28643dc04-kube-api-access-z98n2\") pod \"etcd-operator-b45778765-628ks\" (UID: \"56d0bac1-fb03-49ab-8e45-1cb28643dc04\") " pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.254699 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afb48788-1b96-4ba0-9a9d-98e613451c6f-config\") pod \"kube-apiserver-operator-766d6c64bb-blj6k\" (UID: \"afb48788-1b96-4ba0-9a9d-98e613451c6f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-blj6k" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.254748 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/871099ce-cd86-498e-b227-ec40b6c29e47-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-q8jqv\" (UID: \"871099ce-cd86-498e-b227-ec40b6c29e47\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8jqv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.254776 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc7s6\" (UniqueName: \"kubernetes.io/projected/11782075-8384-415d-a1b9-6c972f8f6f66-kube-api-access-bc7s6\") pod \"control-plane-machine-set-operator-78cbb6b69f-jl78z\" (UID: \"11782075-8384-415d-a1b9-6c972f8f6f66\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jl78z" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.254798 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/871099ce-cd86-498e-b227-ec40b6c29e47-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-q8jqv\" (UID: \"871099ce-cd86-498e-b227-ec40b6c29e47\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8jqv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.254820 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/56d0bac1-fb03-49ab-8e45-1cb28643dc04-etcd-service-ca\") pod \"etcd-operator-b45778765-628ks\" (UID: \"56d0bac1-fb03-49ab-8e45-1cb28643dc04\") " pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.254837 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzq7d\" (UniqueName: \"kubernetes.io/projected/dbc348aa-a6f2-4fee-8c96-83f98c8358b9-kube-api-access-mzq7d\") pod \"migrator-59844c95c7-q95v9\" (UID: \"dbc348aa-a6f2-4fee-8c96-83f98c8358b9\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q95v9" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.254852 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/56d0bac1-fb03-49ab-8e45-1cb28643dc04-etcd-client\") pod \"etcd-operator-b45778765-628ks\" (UID: \"56d0bac1-fb03-49ab-8e45-1cb28643dc04\") " pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.254869 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56d0bac1-fb03-49ab-8e45-1cb28643dc04-config\") pod \"etcd-operator-b45778765-628ks\" (UID: \"56d0bac1-fb03-49ab-8e45-1cb28643dc04\") " pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.254949 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/afb48788-1b96-4ba0-9a9d-98e613451c6f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-blj6k\" (UID: \"afb48788-1b96-4ba0-9a9d-98e613451c6f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-blj6k" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.254965 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d69wz\" (UniqueName: \"kubernetes.io/projected/a58048c2-af3b-4ea3-820e-622c72edd91f-kube-api-access-d69wz\") pod \"cluster-samples-operator-665b6dd947-cn7qm\" (UID: \"a58048c2-af3b-4ea3-820e-622c72edd91f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cn7qm" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.254985 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/11782075-8384-415d-a1b9-6c972f8f6f66-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jl78z\" (UID: \"11782075-8384-415d-a1b9-6c972f8f6f66\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jl78z" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.255002 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/56d0bac1-fb03-49ab-8e45-1cb28643dc04-etcd-ca\") pod \"etcd-operator-b45778765-628ks\" (UID: \"56d0bac1-fb03-49ab-8e45-1cb28643dc04\") " pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.255017 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afb48788-1b96-4ba0-9a9d-98e613451c6f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-blj6k\" (UID: \"afb48788-1b96-4ba0-9a9d-98e613451c6f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-blj6k" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.255053 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a58048c2-af3b-4ea3-820e-622c72edd91f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-cn7qm\" (UID: \"a58048c2-af3b-4ea3-820e-622c72edd91f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cn7qm" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.255066 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56d0bac1-fb03-49ab-8e45-1cb28643dc04-serving-cert\") pod \"etcd-operator-b45778765-628ks\" (UID: \"56d0bac1-fb03-49ab-8e45-1cb28643dc04\") " pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.255084 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lpz9\" (UniqueName: \"kubernetes.io/projected/871099ce-cd86-498e-b227-ec40b6c29e47-kube-api-access-2lpz9\") pod \"openshift-controller-manager-operator-756b6f6bc6-q8jqv\" (UID: \"871099ce-cd86-498e-b227-ec40b6c29e47\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8jqv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.255650 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56d0bac1-fb03-49ab-8e45-1cb28643dc04-config\") pod \"etcd-operator-b45778765-628ks\" (UID: \"56d0bac1-fb03-49ab-8e45-1cb28643dc04\") " pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.255635 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/56d0bac1-fb03-49ab-8e45-1cb28643dc04-etcd-service-ca\") pod \"etcd-operator-b45778765-628ks\" (UID: \"56d0bac1-fb03-49ab-8e45-1cb28643dc04\") " pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.255879 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/56d0bac1-fb03-49ab-8e45-1cb28643dc04-etcd-ca\") pod \"etcd-operator-b45778765-628ks\" (UID: \"56d0bac1-fb03-49ab-8e45-1cb28643dc04\") " pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.258672 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56d0bac1-fb03-49ab-8e45-1cb28643dc04-serving-cert\") pod \"etcd-operator-b45778765-628ks\" (UID: \"56d0bac1-fb03-49ab-8e45-1cb28643dc04\") " pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.258848 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/56d0bac1-fb03-49ab-8e45-1cb28643dc04-etcd-client\") pod \"etcd-operator-b45778765-628ks\" (UID: \"56d0bac1-fb03-49ab-8e45-1cb28643dc04\") " pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.258992 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a58048c2-af3b-4ea3-820e-622c72edd91f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-cn7qm\" (UID: \"a58048c2-af3b-4ea3-820e-622c72edd91f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cn7qm" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.259440 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/11782075-8384-415d-a1b9-6c972f8f6f66-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jl78z\" (UID: \"11782075-8384-415d-a1b9-6c972f8f6f66\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jl78z" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.263384 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.298447 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srr7v\" (UniqueName: \"kubernetes.io/projected/dc6f463d-e632-4eed-88ef-dcd0b7dc0b74-kube-api-access-srr7v\") pod \"apiserver-7bbb656c7d-k2nbx\" (UID: \"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.323062 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.343444 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.365010 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.382558 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.402533 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.423146 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.428950 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/871099ce-cd86-498e-b227-ec40b6c29e47-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-q8jqv\" (UID: \"871099ce-cd86-498e-b227-ec40b6c29e47\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8jqv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.442701 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.446124 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/871099ce-cd86-498e-b227-ec40b6c29e47-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-q8jqv\" (UID: \"871099ce-cd86-498e-b227-ec40b6c29e47\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8jqv" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.463644 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.483343 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.503452 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.523203 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.543589 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.563336 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.565500 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.584508 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.603148 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.623356 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.630769 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2q8lz"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.643290 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.663419 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.683016 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.702888 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.723559 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.743169 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.756520 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx"] Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.762693 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.783060 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.802940 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.823714 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.829102 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afb48788-1b96-4ba0-9a9d-98e613451c6f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-blj6k\" (UID: \"afb48788-1b96-4ba0-9a9d-98e613451c6f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-blj6k" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.842492 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.846376 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afb48788-1b96-4ba0-9a9d-98e613451c6f-config\") pod \"kube-apiserver-operator-766d6c64bb-blj6k\" (UID: \"afb48788-1b96-4ba0-9a9d-98e613451c6f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-blj6k" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.882976 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.903230 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.923714 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.944050 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.963474 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 21 17:58:30 crc kubenswrapper[4808]: I0121 17:58:30.983339 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.002318 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.023066 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.050735 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.063069 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.084262 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.103514 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.123723 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.141215 4808 request.go:700] Waited for 1.008771665s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/secrets?fieldSelector=metadata.name%3Drouter-certs-default&limit=500&resourceVersion=0 Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.144471 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.162899 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.183136 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.203210 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.224067 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.243555 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.262133 4808 generic.go:334] "Generic (PLEG): container finished" podID="dc6f463d-e632-4eed-88ef-dcd0b7dc0b74" containerID="34826a34c238020991ae1b2f4e61f52955162d739c764cd360d80452588b8bbd" exitCode=0 Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.262206 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" event={"ID":"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74","Type":"ContainerDied","Data":"34826a34c238020991ae1b2f4e61f52955162d739c764cd360d80452588b8bbd"} Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.262231 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" event={"ID":"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74","Type":"ContainerStarted","Data":"d371401ec415903d0b111f2e0f67636fd0fc8a1e580863ef9e74eb13358a2109"} Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.262749 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.264811 4808 generic.go:334] "Generic (PLEG): container finished" podID="9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d" containerID="c4150ab05f3e5d7b0a79ee090b15bbf67d7f6c9139fb6138ff920627c9ada682" exitCode=0 Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.264855 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" event={"ID":"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d","Type":"ContainerDied","Data":"c4150ab05f3e5d7b0a79ee090b15bbf67d7f6c9139fb6138ff920627c9ada682"} Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.264886 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" event={"ID":"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d","Type":"ContainerStarted","Data":"8d3401d95d5e85d4357c81b606d8051a7ccfbd77a8b602902f7f522e3a8c055e"} Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.283910 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.303680 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.323393 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.343201 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.363169 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.383611 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.403575 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.423152 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.443334 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.466189 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.482752 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.502592 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.523166 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.544365 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.563129 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.574269 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:31 crc kubenswrapper[4808]: E0121 17:58:31.574789 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 18:00:33.574769003 +0000 UTC m=+268.305279888 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.582791 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.602937 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.622653 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.643161 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.663307 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.675716 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.675808 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.675863 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.676010 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.676855 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.680587 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.681192 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.685596 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.687375 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.702878 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.724815 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.739411 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.742911 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.752294 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.770799 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.775983 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.800306 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-589nd\" (UniqueName: \"kubernetes.io/projected/3919133e-aaba-4dc1-b4cb-6b09f6440221-kube-api-access-589nd\") pod \"controller-manager-879f6c89f-8h66x\" (UID: \"3919133e-aaba-4dc1-b4cb-6b09f6440221\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.827652 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8cb6\" (UniqueName: \"kubernetes.io/projected/05688b08-2c9f-4c6c-967a-ff64de7837c0-kube-api-access-c8cb6\") pod \"console-f9d7485db-h6gm7\" (UID: \"05688b08-2c9f-4c6c-967a-ff64de7837c0\") " pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.843831 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvdcv\" (UniqueName: \"kubernetes.io/projected/a189dfc0-4fe2-4290-b881-f238a055ab5a-kube-api-access-lvdcv\") pod \"console-operator-58897d9998-fc6zr\" (UID: \"a189dfc0-4fe2-4290-b881-f238a055ab5a\") " pod="openshift-console-operator/console-operator-58897d9998-fc6zr" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.858962 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7zt6\" (UniqueName: \"kubernetes.io/projected/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-kube-api-access-x7zt6\") pod \"route-controller-manager-6576b87f9c-wn2m8\" (UID: \"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.878135 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxrs7\" (UniqueName: \"kubernetes.io/projected/1c1e9bbd-24f9-4ef3-8844-6ddb9861c509-kube-api-access-jxrs7\") pod \"openshift-config-operator-7777fb866f-p9d2b\" (UID: \"1c1e9bbd-24f9-4ef3-8844-6ddb9861c509\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.898986 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54slc\" (UniqueName: \"kubernetes.io/projected/ab1aa451-ae2a-48ec-ae13-0689bad3c894-kube-api-access-54slc\") pod \"machine-approver-56656f9798-ps7kv\" (UID: \"ab1aa451-ae2a-48ec-ae13-0689bad3c894\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ps7kv" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.901673 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.902912 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.913651 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-fc6zr" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.925896 4808 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.942824 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.979721 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5c7z\" (UniqueName: \"kubernetes.io/projected/4c87e4f3-6255-4fe3-a52c-726a1acf7ba8-kube-api-access-s5c7z\") pod \"machine-api-operator-5694c8668f-4g7bd\" (UID: \"4c87e4f3-6255-4fe3-a52c-726a1acf7ba8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g7bd" Jan 21 17:58:31 crc kubenswrapper[4808]: I0121 17:58:31.999469 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.002993 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n95rc\" (UniqueName: \"kubernetes.io/projected/5b91a9d5-e508-4624-a43c-770a9dcdbc1d-kube-api-access-n95rc\") pod \"cluster-image-registry-operator-dc59b4c8b-lq8cg\" (UID: \"5b91a9d5-e508-4624-a43c-770a9dcdbc1d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lq8cg" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.029466 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlc27\" (UniqueName: \"kubernetes.io/projected/26022c12-cb37-40ed-9ecb-aef5f82d2bed-kube-api-access-dlc27\") pod \"openshift-apiserver-operator-796bbdcf4f-hr5cr\" (UID: \"26022c12-cb37-40ed-9ecb-aef5f82d2bed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hr5cr" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.051488 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d992\" (UniqueName: \"kubernetes.io/projected/196e1220-375f-4a05-90e0-ce571f606de4-kube-api-access-5d992\") pod \"downloads-7954f5f757-ftpvl\" (UID: \"196e1220-375f-4a05-90e0-ce571f606de4\") " pod="openshift-console/downloads-7954f5f757-ftpvl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.073133 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.080935 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ps7kv" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.089748 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5b91a9d5-e508-4624-a43c-770a9dcdbc1d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lq8cg\" (UID: \"5b91a9d5-e508-4624-a43c-770a9dcdbc1d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lq8cg" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.091317 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.105594 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vzpx\" (UniqueName: \"kubernetes.io/projected/05d34d86-bf40-4b28-94b9-8292fe685637-kube-api-access-7vzpx\") pod \"authentication-operator-69f744f599-jcnqg\" (UID: \"05d34d86-bf40-4b28-94b9-8292fe685637\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jcnqg" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.107705 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.112849 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-4g7bd" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.113748 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skpzt\" (UniqueName: \"kubernetes.io/projected/66c5c3c8-4515-401a-bfe9-5755068bc771-kube-api-access-skpzt\") pod \"oauth-openshift-558db77b4-bz45s\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:32 crc kubenswrapper[4808]: W0121 17:58:32.118226 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-3adddab4ca67e028526d23489d4e519d895aa77b57beb56aee63b6d14d0bf675 WatchSource:0}: Error finding container 3adddab4ca67e028526d23489d4e519d895aa77b57beb56aee63b6d14d0bf675: Status 404 returned error can't find the container with id 3adddab4ca67e028526d23489d4e519d895aa77b57beb56aee63b6d14d0bf675 Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.124699 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 21 17:58:32 crc kubenswrapper[4808]: W0121 17:58:32.132750 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab1aa451_ae2a_48ec_ae13_0689bad3c894.slice/crio-caeedb17942f84bcc288bc57924b1be88d5263947ee9b0bca6256d8be43b458d WatchSource:0}: Error finding container caeedb17942f84bcc288bc57924b1be88d5263947ee9b0bca6256d8be43b458d: Status 404 returned error can't find the container with id caeedb17942f84bcc288bc57924b1be88d5263947ee9b0bca6256d8be43b458d Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.141669 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hr5cr" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.142094 4808 request.go:700] Waited for 1.939159454s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/secrets?fieldSelector=metadata.name%3Ddns-default-metrics-tls&limit=500&resourceVersion=0 Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.143728 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.164001 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.184540 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.202965 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.214531 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8h66x"] Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.223331 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.237032 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.245088 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-ftpvl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.261185 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.262369 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.281335 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"1f92f13afaffdb18feecbe52aa8f64e13615670e59f4cab66a0f32b82cf971ea"} Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.284537 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.285920 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-jcnqg" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.311659 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lq8cg" Jan 21 17:58:32 crc kubenswrapper[4808]: W0121 17:58:32.319733 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3919133e_aaba_4dc1_b4cb_6b09f6440221.slice/crio-0fafb665cfc5ffb6789f9e9c4f7e96b7efe3a1a1c6c27490504d9a8d24447553 WatchSource:0}: Error finding container 0fafb665cfc5ffb6789f9e9c4f7e96b7efe3a1a1c6c27490504d9a8d24447553: Status 404 returned error can't find the container with id 0fafb665cfc5ffb6789f9e9c4f7e96b7efe3a1a1c6c27490504d9a8d24447553 Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.321436 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"97b77c6a4dec6d16a7586e6d6499b46cec7bdbd86c6e6a27b0b6be39339252e0"} Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.321488 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7ef2e942bcdadc99d25d01e3788f18ab01d6c6da635786f1cef368fb8a0302c7"} Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.326421 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z98n2\" (UniqueName: \"kubernetes.io/projected/56d0bac1-fb03-49ab-8e45-1cb28643dc04-kube-api-access-z98n2\") pod \"etcd-operator-b45778765-628ks\" (UID: \"56d0bac1-fb03-49ab-8e45-1cb28643dc04\") " pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.330556 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" event={"ID":"dc6f463d-e632-4eed-88ef-dcd0b7dc0b74","Type":"ContainerStarted","Data":"00f67e3974d523d3112af0812623939877b82ea5d9b501054c17d907d1f9d85d"} Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.338008 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" event={"ID":"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d","Type":"ContainerStarted","Data":"cff421efdad489be5dfb170eedf7e127726ae0d8ba34323b61421ecad7bc9468"} Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.338041 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" event={"ID":"9d028ae7-0f9c-4e2e-b22d-7d4a6aeada9d","Type":"ContainerStarted","Data":"d226faa76c825f8717324db6a0682b841604b43d0fc8ef9c9be8f6d410a0cdd1"} Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.339916 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ps7kv" event={"ID":"ab1aa451-ae2a-48ec-ae13-0689bad3c894","Type":"ContainerStarted","Data":"caeedb17942f84bcc288bc57924b1be88d5263947ee9b0bca6256d8be43b458d"} Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.343650 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc7s6\" (UniqueName: \"kubernetes.io/projected/11782075-8384-415d-a1b9-6c972f8f6f66-kube-api-access-bc7s6\") pod \"control-plane-machine-set-operator-78cbb6b69f-jl78z\" (UID: \"11782075-8384-415d-a1b9-6c972f8f6f66\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jl78z" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.345771 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"3adddab4ca67e028526d23489d4e519d895aa77b57beb56aee63b6d14d0bf675"} Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.363257 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzq7d\" (UniqueName: \"kubernetes.io/projected/dbc348aa-a6f2-4fee-8c96-83f98c8358b9-kube-api-access-mzq7d\") pod \"migrator-59844c95c7-q95v9\" (UID: \"dbc348aa-a6f2-4fee-8c96-83f98c8358b9\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q95v9" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.367929 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q95v9" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.390538 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/afb48788-1b96-4ba0-9a9d-98e613451c6f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-blj6k\" (UID: \"afb48788-1b96-4ba0-9a9d-98e613451c6f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-blj6k" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.395795 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-blj6k" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.412187 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lpz9\" (UniqueName: \"kubernetes.io/projected/871099ce-cd86-498e-b227-ec40b6c29e47-kube-api-access-2lpz9\") pod \"openshift-controller-manager-operator-756b6f6bc6-q8jqv\" (UID: \"871099ce-cd86-498e-b227-ec40b6c29e47\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8jqv" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.420048 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d69wz\" (UniqueName: \"kubernetes.io/projected/a58048c2-af3b-4ea3-820e-622c72edd91f-kube-api-access-d69wz\") pod \"cluster-samples-operator-665b6dd947-cn7qm\" (UID: \"a58048c2-af3b-4ea3-820e-622c72edd91f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cn7qm" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.428206 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-h6gm7"] Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.463445 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-fc6zr"] Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.491483 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b3e892c-c86a-42cd-b257-d722cd033b5e-registry-certificates\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.491938 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e897d5b0-11f6-4e3f-99d2-22a7d093738a-images\") pod \"machine-config-operator-74547568cd-n6tbv\" (UID: \"e897d5b0-11f6-4e3f-99d2-22a7d093738a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.491968 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e897d5b0-11f6-4e3f-99d2-22a7d093738a-proxy-tls\") pod \"machine-config-operator-74547568cd-n6tbv\" (UID: \"e897d5b0-11f6-4e3f-99d2-22a7d093738a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.492031 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/433911de-428a-4261-8680-66a30c308919-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-slth9\" (UID: \"433911de-428a-4261-8680-66a30c308919\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-slth9" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.492094 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqklb\" (UniqueName: \"kubernetes.io/projected/e897d5b0-11f6-4e3f-99d2-22a7d093738a-kube-api-access-hqklb\") pod \"machine-config-operator-74547568cd-n6tbv\" (UID: \"e897d5b0-11f6-4e3f-99d2-22a7d093738a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.492138 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b3e892c-c86a-42cd-b257-d722cd033b5e-bound-sa-token\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.492173 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/030a3f02-a9cd-4a3c-890d-686e2b7df274-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wj246\" (UID: \"030a3f02-a9cd-4a3c-890d-686e2b7df274\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wj246" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.492208 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/030a3f02-a9cd-4a3c-890d-686e2b7df274-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wj246\" (UID: \"030a3f02-a9cd-4a3c-890d-686e2b7df274\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wj246" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.492295 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/030a3f02-a9cd-4a3c-890d-686e2b7df274-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wj246\" (UID: \"030a3f02-a9cd-4a3c-890d-686e2b7df274\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wj246" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.492339 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e897d5b0-11f6-4e3f-99d2-22a7d093738a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-n6tbv\" (UID: \"e897d5b0-11f6-4e3f-99d2-22a7d093738a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.492373 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a848b67a-a4cd-4057-8a2e-2587d6b7142d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-xcn7d\" (UID: \"a848b67a-a4cd-4057-8a2e-2587d6b7142d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xcn7d" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.492411 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/433911de-428a-4261-8680-66a30c308919-config\") pod \"kube-controller-manager-operator-78b949d7b-slth9\" (UID: \"433911de-428a-4261-8680-66a30c308919\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-slth9" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.492436 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a848b67a-a4cd-4057-8a2e-2587d6b7142d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-xcn7d\" (UID: \"a848b67a-a4cd-4057-8a2e-2587d6b7142d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xcn7d" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.492458 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wptxh\" (UniqueName: \"kubernetes.io/projected/2b3e892c-c86a-42cd-b257-d722cd033b5e-kube-api-access-wptxh\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.492506 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b3e892c-c86a-42cd-b257-d722cd033b5e-trusted-ca\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.492531 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b3e892c-c86a-42cd-b257-d722cd033b5e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.492595 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b3e892c-c86a-42cd-b257-d722cd033b5e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.492671 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsg9c\" (UniqueName: \"kubernetes.io/projected/a848b67a-a4cd-4057-8a2e-2587d6b7142d-kube-api-access-qsg9c\") pod \"kube-storage-version-migrator-operator-b67b599dd-xcn7d\" (UID: \"a848b67a-a4cd-4057-8a2e-2587d6b7142d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xcn7d" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.493368 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b3e892c-c86a-42cd-b257-d722cd033b5e-registry-tls\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.495008 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.495095 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/433911de-428a-4261-8680-66a30c308919-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-slth9\" (UID: \"433911de-428a-4261-8680-66a30c308919\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-slth9" Jan 21 17:58:32 crc kubenswrapper[4808]: E0121 17:58:32.495389 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:32.995376149 +0000 UTC m=+147.725887044 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.595887 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:32 crc kubenswrapper[4808]: E0121 17:58:32.596042 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:33.096020393 +0000 UTC m=+147.826531278 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596100 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/77ae37d7-ba3c-4b77-85cb-f7cb648ab795-apiservice-cert\") pod \"packageserver-d55dfcdfc-qfggh\" (UID: \"77ae37d7-ba3c-4b77-85cb-f7cb648ab795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596168 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/044bc3f1-737c-4dda-9c24-d087ce039b42-config-volume\") pod \"dns-default-zjcb6\" (UID: \"044bc3f1-737c-4dda-9c24-d087ce039b42\") " pod="openshift-dns/dns-default-zjcb6" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596185 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rdrb\" (UniqueName: \"kubernetes.io/projected/c6350524-fc0a-4bbc-b803-cea5fd8dc76c-kube-api-access-8rdrb\") pod \"olm-operator-6b444d44fb-4gg7j\" (UID: \"c6350524-fc0a-4bbc-b803-cea5fd8dc76c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4gg7j" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596225 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fx7n\" (UniqueName: \"kubernetes.io/projected/922aad90-2298-4afb-816e-063c73f18c6d-kube-api-access-7fx7n\") pod \"csi-hostpathplugin-glhp5\" (UID: \"922aad90-2298-4afb-816e-063c73f18c6d\") " pod="hostpath-provisioner/csi-hostpathplugin-glhp5" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596303 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgd64\" (UniqueName: \"kubernetes.io/projected/c0b6ae2d-bfaf-4535-9906-31b19d4fe489-kube-api-access-bgd64\") pod \"collect-profiles-29483625-lsqb2\" (UID: \"c0b6ae2d-bfaf-4535-9906-31b19d4fe489\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596346 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596364 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b5ad8de7-53f2-475d-be7b-acb85ad7f53a-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-jxgs8\" (UID: \"b5ad8de7-53f2-475d-be7b-acb85ad7f53a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jxgs8" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596387 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh256\" (UniqueName: \"kubernetes.io/projected/2c1a10ce-320b-48fe-9eca-ab5e6d110441-kube-api-access-nh256\") pod \"catalog-operator-68c6474976-jmp6c\" (UID: \"2c1a10ce-320b-48fe-9eca-ab5e6d110441\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jmp6c" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596412 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b3e892c-c86a-42cd-b257-d722cd033b5e-registry-certificates\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596436 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-pvk7n\" (UID: \"5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5\") " pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596476 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m679v\" (UniqueName: \"kubernetes.io/projected/d98f2d52-51f0-4843-8ba3-50264b3b3abe-kube-api-access-m679v\") pod \"ingress-operator-5b745b69d9-mqm6k\" (UID: \"d98f2d52-51f0-4843-8ba3-50264b3b3abe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596499 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c6350524-fc0a-4bbc-b803-cea5fd8dc76c-srv-cert\") pod \"olm-operator-6b444d44fb-4gg7j\" (UID: \"c6350524-fc0a-4bbc-b803-cea5fd8dc76c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4gg7j" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596518 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/044bc3f1-737c-4dda-9c24-d087ce039b42-metrics-tls\") pod \"dns-default-zjcb6\" (UID: \"044bc3f1-737c-4dda-9c24-d087ce039b42\") " pod="openshift-dns/dns-default-zjcb6" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596554 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e897d5b0-11f6-4e3f-99d2-22a7d093738a-images\") pod \"machine-config-operator-74547568cd-n6tbv\" (UID: \"e897d5b0-11f6-4e3f-99d2-22a7d093738a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596629 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9a23bae9-c21c-40a6-aa8d-9c3524614dd4-default-certificate\") pod \"router-default-5444994796-xfblx\" (UID: \"9a23bae9-c21c-40a6-aa8d-9c3524614dd4\") " pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596668 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/433911de-428a-4261-8680-66a30c308919-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-slth9\" (UID: \"433911de-428a-4261-8680-66a30c308919\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-slth9" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596739 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqklb\" (UniqueName: \"kubernetes.io/projected/e897d5b0-11f6-4e3f-99d2-22a7d093738a-kube-api-access-hqklb\") pod \"machine-config-operator-74547568cd-n6tbv\" (UID: \"e897d5b0-11f6-4e3f-99d2-22a7d093738a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596769 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2c1a10ce-320b-48fe-9eca-ab5e6d110441-srv-cert\") pod \"catalog-operator-68c6474976-jmp6c\" (UID: \"2c1a10ce-320b-48fe-9eca-ab5e6d110441\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jmp6c" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596791 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hwlx\" (UniqueName: \"kubernetes.io/projected/e6e2b28f-7cfe-44a5-8a63-f17542f27cb6-kube-api-access-6hwlx\") pod \"service-ca-9c57cc56f-s5khx\" (UID: \"e6e2b28f-7cfe-44a5-8a63-f17542f27cb6\") " pod="openshift-service-ca/service-ca-9c57cc56f-s5khx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596807 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a23bae9-c21c-40a6-aa8d-9c3524614dd4-service-ca-bundle\") pod \"router-default-5444994796-xfblx\" (UID: \"9a23bae9-c21c-40a6-aa8d-9c3524614dd4\") " pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596822 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0b6ae2d-bfaf-4535-9906-31b19d4fe489-secret-volume\") pod \"collect-profiles-29483625-lsqb2\" (UID: \"c0b6ae2d-bfaf-4535-9906-31b19d4fe489\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596878 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/030a3f02-a9cd-4a3c-890d-686e2b7df274-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wj246\" (UID: \"030a3f02-a9cd-4a3c-890d-686e2b7df274\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wj246" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596956 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e897d5b0-11f6-4e3f-99d2-22a7d093738a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-n6tbv\" (UID: \"e897d5b0-11f6-4e3f-99d2-22a7d093738a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.596993 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wptxh\" (UniqueName: \"kubernetes.io/projected/2b3e892c-c86a-42cd-b257-d722cd033b5e-kube-api-access-wptxh\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597015 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/77ae37d7-ba3c-4b77-85cb-f7cb648ab795-tmpfs\") pod \"packageserver-d55dfcdfc-qfggh\" (UID: \"77ae37d7-ba3c-4b77-85cb-f7cb648ab795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597085 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c8a01e69-a842-4316-9fd6-a2850afa5b11-proxy-tls\") pod \"machine-config-controller-84d6567774-d2rkk\" (UID: \"c8a01e69-a842-4316-9fd6-a2850afa5b11\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d2rkk" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597125 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmq7q\" (UniqueName: \"kubernetes.io/projected/d9f360c8-31a2-4577-8eec-d5d58c106933-kube-api-access-gmq7q\") pod \"dns-operator-744455d44c-z77hq\" (UID: \"d9f360c8-31a2-4577-8eec-d5d58c106933\") " pod="openshift-dns-operator/dns-operator-744455d44c-z77hq" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597154 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d9f360c8-31a2-4577-8eec-d5d58c106933-metrics-tls\") pod \"dns-operator-744455d44c-z77hq\" (UID: \"d9f360c8-31a2-4577-8eec-d5d58c106933\") " pod="openshift-dns-operator/dns-operator-744455d44c-z77hq" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597188 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c8a01e69-a842-4316-9fd6-a2850afa5b11-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-d2rkk\" (UID: \"c8a01e69-a842-4316-9fd6-a2850afa5b11\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d2rkk" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597203 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/922aad90-2298-4afb-816e-063c73f18c6d-socket-dir\") pod \"csi-hostpathplugin-glhp5\" (UID: \"922aad90-2298-4afb-816e-063c73f18c6d\") " pod="hostpath-provisioner/csi-hostpathplugin-glhp5" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597262 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f669bbf5-0b37-4821-9b15-052eef33004e-node-bootstrap-token\") pod \"machine-config-server-2m5jx\" (UID: \"f669bbf5-0b37-4821-9b15-052eef33004e\") " pod="openshift-machine-config-operator/machine-config-server-2m5jx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597280 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp5kl\" (UniqueName: \"kubernetes.io/projected/bc0b3e27-43a2-4a5b-98bc-d371958b2bbf-kube-api-access-jp5kl\") pod \"package-server-manager-789f6589d5-q8wfq\" (UID: \"bc0b3e27-43a2-4a5b-98bc-d371958b2bbf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q8wfq" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597314 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j99pq\" (UniqueName: \"kubernetes.io/projected/77ae37d7-ba3c-4b77-85cb-f7cb648ab795-kube-api-access-j99pq\") pod \"packageserver-d55dfcdfc-qfggh\" (UID: \"77ae37d7-ba3c-4b77-85cb-f7cb648ab795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597329 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c71840-39bc-4256-bdfb-43d6e1bb749e-config\") pod \"service-ca-operator-777779d784-fn8s9\" (UID: \"99c71840-39bc-4256-bdfb-43d6e1bb749e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fn8s9" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597373 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b3e892c-c86a-42cd-b257-d722cd033b5e-registry-tls\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597388 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/922aad90-2298-4afb-816e-063c73f18c6d-registration-dir\") pod \"csi-hostpathplugin-glhp5\" (UID: \"922aad90-2298-4afb-816e-063c73f18c6d\") " pod="hostpath-provisioner/csi-hostpathplugin-glhp5" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597414 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27k7s\" (UniqueName: \"kubernetes.io/projected/29185a07-7266-4975-870b-127f880f2cf6-kube-api-access-27k7s\") pod \"ingress-canary-79g2n\" (UID: \"29185a07-7266-4975-870b-127f880f2cf6\") " pod="openshift-ingress-canary/ingress-canary-79g2n" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597430 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e6e2b28f-7cfe-44a5-8a63-f17542f27cb6-signing-cabundle\") pod \"service-ca-9c57cc56f-s5khx\" (UID: \"e6e2b28f-7cfe-44a5-8a63-f17542f27cb6\") " pod="openshift-service-ca/service-ca-9c57cc56f-s5khx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597501 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9fd6\" (UniqueName: \"kubernetes.io/projected/f669bbf5-0b37-4821-9b15-052eef33004e-kube-api-access-m9fd6\") pod \"machine-config-server-2m5jx\" (UID: \"f669bbf5-0b37-4821-9b15-052eef33004e\") " pod="openshift-machine-config-operator/machine-config-server-2m5jx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597528 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/433911de-428a-4261-8680-66a30c308919-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-slth9\" (UID: \"433911de-428a-4261-8680-66a30c308919\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-slth9" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597582 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29185a07-7266-4975-870b-127f880f2cf6-cert\") pod \"ingress-canary-79g2n\" (UID: \"29185a07-7266-4975-870b-127f880f2cf6\") " pod="openshift-ingress-canary/ingress-canary-79g2n" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597625 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5dpk\" (UniqueName: \"kubernetes.io/projected/044bc3f1-737c-4dda-9c24-d087ce039b42-kube-api-access-j5dpk\") pod \"dns-default-zjcb6\" (UID: \"044bc3f1-737c-4dda-9c24-d087ce039b42\") " pod="openshift-dns/dns-default-zjcb6" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597672 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0b6ae2d-bfaf-4535-9906-31b19d4fe489-config-volume\") pod \"collect-profiles-29483625-lsqb2\" (UID: \"c0b6ae2d-bfaf-4535-9906-31b19d4fe489\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597699 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/bc0b3e27-43a2-4a5b-98bc-d371958b2bbf-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-q8wfq\" (UID: \"bc0b3e27-43a2-4a5b-98bc-d371958b2bbf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q8wfq" Jan 21 17:58:32 crc kubenswrapper[4808]: E0121 17:58:32.597737 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:33.097720513 +0000 UTC m=+147.828231398 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597815 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjphr\" (UniqueName: \"kubernetes.io/projected/b5ad8de7-53f2-475d-be7b-acb85ad7f53a-kube-api-access-gjphr\") pod \"multus-admission-controller-857f4d67dd-jxgs8\" (UID: \"b5ad8de7-53f2-475d-be7b-acb85ad7f53a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jxgs8" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597837 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a23bae9-c21c-40a6-aa8d-9c3524614dd4-metrics-certs\") pod \"router-default-5444994796-xfblx\" (UID: \"9a23bae9-c21c-40a6-aa8d-9c3524614dd4\") " pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597873 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e897d5b0-11f6-4e3f-99d2-22a7d093738a-proxy-tls\") pod \"machine-config-operator-74547568cd-n6tbv\" (UID: \"e897d5b0-11f6-4e3f-99d2-22a7d093738a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597920 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e6e2b28f-7cfe-44a5-8a63-f17542f27cb6-signing-key\") pod \"service-ca-9c57cc56f-s5khx\" (UID: \"e6e2b28f-7cfe-44a5-8a63-f17542f27cb6\") " pod="openshift-service-ca/service-ca-9c57cc56f-s5khx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597946 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/922aad90-2298-4afb-816e-063c73f18c6d-csi-data-dir\") pod \"csi-hostpathplugin-glhp5\" (UID: \"922aad90-2298-4afb-816e-063c73f18c6d\") " pod="hostpath-provisioner/csi-hostpathplugin-glhp5" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.597970 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d98f2d52-51f0-4843-8ba3-50264b3b3abe-trusted-ca\") pod \"ingress-operator-5b745b69d9-mqm6k\" (UID: \"d98f2d52-51f0-4843-8ba3-50264b3b3abe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598020 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b3e892c-c86a-42cd-b257-d722cd033b5e-bound-sa-token\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598046 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/030a3f02-a9cd-4a3c-890d-686e2b7df274-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wj246\" (UID: \"030a3f02-a9cd-4a3c-890d-686e2b7df274\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wj246" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598071 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/030a3f02-a9cd-4a3c-890d-686e2b7df274-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wj246\" (UID: \"030a3f02-a9cd-4a3c-890d-686e2b7df274\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wj246" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598097 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d98f2d52-51f0-4843-8ba3-50264b3b3abe-metrics-tls\") pod \"ingress-operator-5b745b69d9-mqm6k\" (UID: \"d98f2d52-51f0-4843-8ba3-50264b3b3abe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598155 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7prvs\" (UniqueName: \"kubernetes.io/projected/5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5-kube-api-access-7prvs\") pod \"marketplace-operator-79b997595-pvk7n\" (UID: \"5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5\") " pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598260 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9a23bae9-c21c-40a6-aa8d-9c3524614dd4-stats-auth\") pod \"router-default-5444994796-xfblx\" (UID: \"9a23bae9-c21c-40a6-aa8d-9c3524614dd4\") " pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598330 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a848b67a-a4cd-4057-8a2e-2587d6b7142d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-xcn7d\" (UID: \"a848b67a-a4cd-4057-8a2e-2587d6b7142d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xcn7d" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598365 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/433911de-428a-4261-8680-66a30c308919-config\") pod \"kube-controller-manager-operator-78b949d7b-slth9\" (UID: \"433911de-428a-4261-8680-66a30c308919\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-slth9" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598384 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/922aad90-2298-4afb-816e-063c73f18c6d-mountpoint-dir\") pod \"csi-hostpathplugin-glhp5\" (UID: \"922aad90-2298-4afb-816e-063c73f18c6d\") " pod="hostpath-provisioner/csi-hostpathplugin-glhp5" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598397 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b3e892c-c86a-42cd-b257-d722cd033b5e-registry-certificates\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598400 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glmg8\" (UniqueName: \"kubernetes.io/projected/99c71840-39bc-4256-bdfb-43d6e1bb749e-kube-api-access-glmg8\") pod \"service-ca-operator-777779d784-fn8s9\" (UID: \"99c71840-39bc-4256-bdfb-43d6e1bb749e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fn8s9" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598449 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-pvk7n\" (UID: \"5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5\") " pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598493 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a848b67a-a4cd-4057-8a2e-2587d6b7142d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-xcn7d\" (UID: \"a848b67a-a4cd-4057-8a2e-2587d6b7142d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xcn7d" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598533 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr6rg\" (UniqueName: \"kubernetes.io/projected/c8a01e69-a842-4316-9fd6-a2850afa5b11-kube-api-access-mr6rg\") pod \"machine-config-controller-84d6567774-d2rkk\" (UID: \"c8a01e69-a842-4316-9fd6-a2850afa5b11\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d2rkk" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598554 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqc87\" (UniqueName: \"kubernetes.io/projected/9a23bae9-c21c-40a6-aa8d-9c3524614dd4-kube-api-access-zqc87\") pod \"router-default-5444994796-xfblx\" (UID: \"9a23bae9-c21c-40a6-aa8d-9c3524614dd4\") " pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598605 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b3e892c-c86a-42cd-b257-d722cd033b5e-trusted-ca\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598649 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b3e892c-c86a-42cd-b257-d722cd033b5e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598713 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/77ae37d7-ba3c-4b77-85cb-f7cb648ab795-webhook-cert\") pod \"packageserver-d55dfcdfc-qfggh\" (UID: \"77ae37d7-ba3c-4b77-85cb-f7cb648ab795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598748 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f669bbf5-0b37-4821-9b15-052eef33004e-certs\") pod \"machine-config-server-2m5jx\" (UID: \"f669bbf5-0b37-4821-9b15-052eef33004e\") " pod="openshift-machine-config-operator/machine-config-server-2m5jx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598778 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c6350524-fc0a-4bbc-b803-cea5fd8dc76c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4gg7j\" (UID: \"c6350524-fc0a-4bbc-b803-cea5fd8dc76c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4gg7j" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598811 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b3e892c-c86a-42cd-b257-d722cd033b5e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598840 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2c1a10ce-320b-48fe-9eca-ab5e6d110441-profile-collector-cert\") pod \"catalog-operator-68c6474976-jmp6c\" (UID: \"2c1a10ce-320b-48fe-9eca-ab5e6d110441\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jmp6c" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598874 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99c71840-39bc-4256-bdfb-43d6e1bb749e-serving-cert\") pod \"service-ca-operator-777779d784-fn8s9\" (UID: \"99c71840-39bc-4256-bdfb-43d6e1bb749e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fn8s9" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598919 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsg9c\" (UniqueName: \"kubernetes.io/projected/a848b67a-a4cd-4057-8a2e-2587d6b7142d-kube-api-access-qsg9c\") pod \"kube-storage-version-migrator-operator-b67b599dd-xcn7d\" (UID: \"a848b67a-a4cd-4057-8a2e-2587d6b7142d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xcn7d" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598950 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/922aad90-2298-4afb-816e-063c73f18c6d-plugins-dir\") pod \"csi-hostpathplugin-glhp5\" (UID: \"922aad90-2298-4afb-816e-063c73f18c6d\") " pod="hostpath-provisioner/csi-hostpathplugin-glhp5" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.598982 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d98f2d52-51f0-4843-8ba3-50264b3b3abe-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mqm6k\" (UID: \"d98f2d52-51f0-4843-8ba3-50264b3b3abe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.600552 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b3e892c-c86a-42cd-b257-d722cd033b5e-trusted-ca\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.603456 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e897d5b0-11f6-4e3f-99d2-22a7d093738a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-n6tbv\" (UID: \"e897d5b0-11f6-4e3f-99d2-22a7d093738a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.604056 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b3e892c-c86a-42cd-b257-d722cd033b5e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.605567 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e897d5b0-11f6-4e3f-99d2-22a7d093738a-images\") pod \"machine-config-operator-74547568cd-n6tbv\" (UID: \"e897d5b0-11f6-4e3f-99d2-22a7d093738a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.607198 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/433911de-428a-4261-8680-66a30c308919-config\") pod \"kube-controller-manager-operator-78b949d7b-slth9\" (UID: \"433911de-428a-4261-8680-66a30c308919\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-slth9" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.607937 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a848b67a-a4cd-4057-8a2e-2587d6b7142d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-xcn7d\" (UID: \"a848b67a-a4cd-4057-8a2e-2587d6b7142d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xcn7d" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.609147 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cn7qm" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.601046 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/030a3f02-a9cd-4a3c-890d-686e2b7df274-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wj246\" (UID: \"030a3f02-a9cd-4a3c-890d-686e2b7df274\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wj246" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.619115 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.627895 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jl78z" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.630430 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a848b67a-a4cd-4057-8a2e-2587d6b7142d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-xcn7d\" (UID: \"a848b67a-a4cd-4057-8a2e-2587d6b7142d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xcn7d" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.635003 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/030a3f02-a9cd-4a3c-890d-686e2b7df274-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wj246\" (UID: \"030a3f02-a9cd-4a3c-890d-686e2b7df274\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wj246" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.636760 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b3e892c-c86a-42cd-b257-d722cd033b5e-registry-tls\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.637651 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/433911de-428a-4261-8680-66a30c308919-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-slth9\" (UID: \"433911de-428a-4261-8680-66a30c308919\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-slth9" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.638014 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e897d5b0-11f6-4e3f-99d2-22a7d093738a-proxy-tls\") pod \"machine-config-operator-74547568cd-n6tbv\" (UID: \"e897d5b0-11f6-4e3f-99d2-22a7d093738a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.640646 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b3e892c-c86a-42cd-b257-d722cd033b5e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.641066 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/433911de-428a-4261-8680-66a30c308919-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-slth9\" (UID: \"433911de-428a-4261-8680-66a30c308919\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-slth9" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.647823 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8jqv" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.657604 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsg9c\" (UniqueName: \"kubernetes.io/projected/a848b67a-a4cd-4057-8a2e-2587d6b7142d-kube-api-access-qsg9c\") pod \"kube-storage-version-migrator-operator-b67b599dd-xcn7d\" (UID: \"a848b67a-a4cd-4057-8a2e-2587d6b7142d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xcn7d" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.678891 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b3e892c-c86a-42cd-b257-d722cd033b5e-bound-sa-token\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.699828 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:32 crc kubenswrapper[4808]: E0121 17:58:32.700583 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:33.200563439 +0000 UTC m=+147.931074314 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702346 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e6e2b28f-7cfe-44a5-8a63-f17542f27cb6-signing-key\") pod \"service-ca-9c57cc56f-s5khx\" (UID: \"e6e2b28f-7cfe-44a5-8a63-f17542f27cb6\") " pod="openshift-service-ca/service-ca-9c57cc56f-s5khx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702373 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/922aad90-2298-4afb-816e-063c73f18c6d-csi-data-dir\") pod \"csi-hostpathplugin-glhp5\" (UID: \"922aad90-2298-4afb-816e-063c73f18c6d\") " pod="hostpath-provisioner/csi-hostpathplugin-glhp5" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702397 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d98f2d52-51f0-4843-8ba3-50264b3b3abe-metrics-tls\") pod \"ingress-operator-5b745b69d9-mqm6k\" (UID: \"d98f2d52-51f0-4843-8ba3-50264b3b3abe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702428 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d98f2d52-51f0-4843-8ba3-50264b3b3abe-trusted-ca\") pod \"ingress-operator-5b745b69d9-mqm6k\" (UID: \"d98f2d52-51f0-4843-8ba3-50264b3b3abe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702444 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7prvs\" (UniqueName: \"kubernetes.io/projected/5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5-kube-api-access-7prvs\") pod \"marketplace-operator-79b997595-pvk7n\" (UID: \"5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5\") " pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702473 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9a23bae9-c21c-40a6-aa8d-9c3524614dd4-stats-auth\") pod \"router-default-5444994796-xfblx\" (UID: \"9a23bae9-c21c-40a6-aa8d-9c3524614dd4\") " pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702512 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/922aad90-2298-4afb-816e-063c73f18c6d-mountpoint-dir\") pod \"csi-hostpathplugin-glhp5\" (UID: \"922aad90-2298-4afb-816e-063c73f18c6d\") " pod="hostpath-provisioner/csi-hostpathplugin-glhp5" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702533 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glmg8\" (UniqueName: \"kubernetes.io/projected/99c71840-39bc-4256-bdfb-43d6e1bb749e-kube-api-access-glmg8\") pod \"service-ca-operator-777779d784-fn8s9\" (UID: \"99c71840-39bc-4256-bdfb-43d6e1bb749e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fn8s9" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702551 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-pvk7n\" (UID: \"5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5\") " pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702595 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mr6rg\" (UniqueName: \"kubernetes.io/projected/c8a01e69-a842-4316-9fd6-a2850afa5b11-kube-api-access-mr6rg\") pod \"machine-config-controller-84d6567774-d2rkk\" (UID: \"c8a01e69-a842-4316-9fd6-a2850afa5b11\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d2rkk" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702615 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqc87\" (UniqueName: \"kubernetes.io/projected/9a23bae9-c21c-40a6-aa8d-9c3524614dd4-kube-api-access-zqc87\") pod \"router-default-5444994796-xfblx\" (UID: \"9a23bae9-c21c-40a6-aa8d-9c3524614dd4\") " pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702663 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/77ae37d7-ba3c-4b77-85cb-f7cb648ab795-webhook-cert\") pod \"packageserver-d55dfcdfc-qfggh\" (UID: \"77ae37d7-ba3c-4b77-85cb-f7cb648ab795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702683 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f669bbf5-0b37-4821-9b15-052eef33004e-certs\") pod \"machine-config-server-2m5jx\" (UID: \"f669bbf5-0b37-4821-9b15-052eef33004e\") " pod="openshift-machine-config-operator/machine-config-server-2m5jx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702705 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c6350524-fc0a-4bbc-b803-cea5fd8dc76c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4gg7j\" (UID: \"c6350524-fc0a-4bbc-b803-cea5fd8dc76c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4gg7j" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702745 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2c1a10ce-320b-48fe-9eca-ab5e6d110441-profile-collector-cert\") pod \"catalog-operator-68c6474976-jmp6c\" (UID: \"2c1a10ce-320b-48fe-9eca-ab5e6d110441\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jmp6c" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702774 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99c71840-39bc-4256-bdfb-43d6e1bb749e-serving-cert\") pod \"service-ca-operator-777779d784-fn8s9\" (UID: \"99c71840-39bc-4256-bdfb-43d6e1bb749e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fn8s9" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702796 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/922aad90-2298-4afb-816e-063c73f18c6d-plugins-dir\") pod \"csi-hostpathplugin-glhp5\" (UID: \"922aad90-2298-4afb-816e-063c73f18c6d\") " pod="hostpath-provisioner/csi-hostpathplugin-glhp5" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702835 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d98f2d52-51f0-4843-8ba3-50264b3b3abe-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mqm6k\" (UID: \"d98f2d52-51f0-4843-8ba3-50264b3b3abe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702858 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/77ae37d7-ba3c-4b77-85cb-f7cb648ab795-apiservice-cert\") pod \"packageserver-d55dfcdfc-qfggh\" (UID: \"77ae37d7-ba3c-4b77-85cb-f7cb648ab795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702899 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/044bc3f1-737c-4dda-9c24-d087ce039b42-config-volume\") pod \"dns-default-zjcb6\" (UID: \"044bc3f1-737c-4dda-9c24-d087ce039b42\") " pod="openshift-dns/dns-default-zjcb6" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702922 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgd64\" (UniqueName: \"kubernetes.io/projected/c0b6ae2d-bfaf-4535-9906-31b19d4fe489-kube-api-access-bgd64\") pod \"collect-profiles-29483625-lsqb2\" (UID: \"c0b6ae2d-bfaf-4535-9906-31b19d4fe489\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702940 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rdrb\" (UniqueName: \"kubernetes.io/projected/c6350524-fc0a-4bbc-b803-cea5fd8dc76c-kube-api-access-8rdrb\") pod \"olm-operator-6b444d44fb-4gg7j\" (UID: \"c6350524-fc0a-4bbc-b803-cea5fd8dc76c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4gg7j" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702955 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fx7n\" (UniqueName: \"kubernetes.io/projected/922aad90-2298-4afb-816e-063c73f18c6d-kube-api-access-7fx7n\") pod \"csi-hostpathplugin-glhp5\" (UID: \"922aad90-2298-4afb-816e-063c73f18c6d\") " pod="hostpath-provisioner/csi-hostpathplugin-glhp5" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.702998 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703016 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b5ad8de7-53f2-475d-be7b-acb85ad7f53a-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-jxgs8\" (UID: \"b5ad8de7-53f2-475d-be7b-acb85ad7f53a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jxgs8" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703030 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh256\" (UniqueName: \"kubernetes.io/projected/2c1a10ce-320b-48fe-9eca-ab5e6d110441-kube-api-access-nh256\") pod \"catalog-operator-68c6474976-jmp6c\" (UID: \"2c1a10ce-320b-48fe-9eca-ab5e6d110441\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jmp6c" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703065 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-pvk7n\" (UID: \"5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5\") " pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703087 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m679v\" (UniqueName: \"kubernetes.io/projected/d98f2d52-51f0-4843-8ba3-50264b3b3abe-kube-api-access-m679v\") pod \"ingress-operator-5b745b69d9-mqm6k\" (UID: \"d98f2d52-51f0-4843-8ba3-50264b3b3abe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703110 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c6350524-fc0a-4bbc-b803-cea5fd8dc76c-srv-cert\") pod \"olm-operator-6b444d44fb-4gg7j\" (UID: \"c6350524-fc0a-4bbc-b803-cea5fd8dc76c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4gg7j" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703155 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/044bc3f1-737c-4dda-9c24-d087ce039b42-metrics-tls\") pod \"dns-default-zjcb6\" (UID: \"044bc3f1-737c-4dda-9c24-d087ce039b42\") " pod="openshift-dns/dns-default-zjcb6" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703180 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9a23bae9-c21c-40a6-aa8d-9c3524614dd4-default-certificate\") pod \"router-default-5444994796-xfblx\" (UID: \"9a23bae9-c21c-40a6-aa8d-9c3524614dd4\") " pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703221 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2c1a10ce-320b-48fe-9eca-ab5e6d110441-srv-cert\") pod \"catalog-operator-68c6474976-jmp6c\" (UID: \"2c1a10ce-320b-48fe-9eca-ab5e6d110441\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jmp6c" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703269 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hwlx\" (UniqueName: \"kubernetes.io/projected/e6e2b28f-7cfe-44a5-8a63-f17542f27cb6-kube-api-access-6hwlx\") pod \"service-ca-9c57cc56f-s5khx\" (UID: \"e6e2b28f-7cfe-44a5-8a63-f17542f27cb6\") " pod="openshift-service-ca/service-ca-9c57cc56f-s5khx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703289 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a23bae9-c21c-40a6-aa8d-9c3524614dd4-service-ca-bundle\") pod \"router-default-5444994796-xfblx\" (UID: \"9a23bae9-c21c-40a6-aa8d-9c3524614dd4\") " pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703307 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0b6ae2d-bfaf-4535-9906-31b19d4fe489-secret-volume\") pod \"collect-profiles-29483625-lsqb2\" (UID: \"c0b6ae2d-bfaf-4535-9906-31b19d4fe489\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703355 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/77ae37d7-ba3c-4b77-85cb-f7cb648ab795-tmpfs\") pod \"packageserver-d55dfcdfc-qfggh\" (UID: \"77ae37d7-ba3c-4b77-85cb-f7cb648ab795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703373 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c8a01e69-a842-4316-9fd6-a2850afa5b11-proxy-tls\") pod \"machine-config-controller-84d6567774-d2rkk\" (UID: \"c8a01e69-a842-4316-9fd6-a2850afa5b11\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d2rkk" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703390 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmq7q\" (UniqueName: \"kubernetes.io/projected/d9f360c8-31a2-4577-8eec-d5d58c106933-kube-api-access-gmq7q\") pod \"dns-operator-744455d44c-z77hq\" (UID: \"d9f360c8-31a2-4577-8eec-d5d58c106933\") " pod="openshift-dns-operator/dns-operator-744455d44c-z77hq" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703407 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d9f360c8-31a2-4577-8eec-d5d58c106933-metrics-tls\") pod \"dns-operator-744455d44c-z77hq\" (UID: \"d9f360c8-31a2-4577-8eec-d5d58c106933\") " pod="openshift-dns-operator/dns-operator-744455d44c-z77hq" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703444 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c8a01e69-a842-4316-9fd6-a2850afa5b11-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-d2rkk\" (UID: \"c8a01e69-a842-4316-9fd6-a2850afa5b11\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d2rkk" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703459 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/922aad90-2298-4afb-816e-063c73f18c6d-socket-dir\") pod \"csi-hostpathplugin-glhp5\" (UID: \"922aad90-2298-4afb-816e-063c73f18c6d\") " pod="hostpath-provisioner/csi-hostpathplugin-glhp5" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703475 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f669bbf5-0b37-4821-9b15-052eef33004e-node-bootstrap-token\") pod \"machine-config-server-2m5jx\" (UID: \"f669bbf5-0b37-4821-9b15-052eef33004e\") " pod="openshift-machine-config-operator/machine-config-server-2m5jx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703519 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp5kl\" (UniqueName: \"kubernetes.io/projected/bc0b3e27-43a2-4a5b-98bc-d371958b2bbf-kube-api-access-jp5kl\") pod \"package-server-manager-789f6589d5-q8wfq\" (UID: \"bc0b3e27-43a2-4a5b-98bc-d371958b2bbf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q8wfq" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703540 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j99pq\" (UniqueName: \"kubernetes.io/projected/77ae37d7-ba3c-4b77-85cb-f7cb648ab795-kube-api-access-j99pq\") pod \"packageserver-d55dfcdfc-qfggh\" (UID: \"77ae37d7-ba3c-4b77-85cb-f7cb648ab795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703558 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c71840-39bc-4256-bdfb-43d6e1bb749e-config\") pod \"service-ca-operator-777779d784-fn8s9\" (UID: \"99c71840-39bc-4256-bdfb-43d6e1bb749e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fn8s9" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703610 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/922aad90-2298-4afb-816e-063c73f18c6d-registration-dir\") pod \"csi-hostpathplugin-glhp5\" (UID: \"922aad90-2298-4afb-816e-063c73f18c6d\") " pod="hostpath-provisioner/csi-hostpathplugin-glhp5" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.703631 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e6e2b28f-7cfe-44a5-8a63-f17542f27cb6-signing-cabundle\") pod \"service-ca-9c57cc56f-s5khx\" (UID: \"e6e2b28f-7cfe-44a5-8a63-f17542f27cb6\") " pod="openshift-service-ca/service-ca-9c57cc56f-s5khx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.705108 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27k7s\" (UniqueName: \"kubernetes.io/projected/29185a07-7266-4975-870b-127f880f2cf6-kube-api-access-27k7s\") pod \"ingress-canary-79g2n\" (UID: \"29185a07-7266-4975-870b-127f880f2cf6\") " pod="openshift-ingress-canary/ingress-canary-79g2n" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.705136 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9fd6\" (UniqueName: \"kubernetes.io/projected/f669bbf5-0b37-4821-9b15-052eef33004e-kube-api-access-m9fd6\") pod \"machine-config-server-2m5jx\" (UID: \"f669bbf5-0b37-4821-9b15-052eef33004e\") " pod="openshift-machine-config-operator/machine-config-server-2m5jx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.705172 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29185a07-7266-4975-870b-127f880f2cf6-cert\") pod \"ingress-canary-79g2n\" (UID: \"29185a07-7266-4975-870b-127f880f2cf6\") " pod="openshift-ingress-canary/ingress-canary-79g2n" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.705188 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5dpk\" (UniqueName: \"kubernetes.io/projected/044bc3f1-737c-4dda-9c24-d087ce039b42-kube-api-access-j5dpk\") pod \"dns-default-zjcb6\" (UID: \"044bc3f1-737c-4dda-9c24-d087ce039b42\") " pod="openshift-dns/dns-default-zjcb6" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.705203 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0b6ae2d-bfaf-4535-9906-31b19d4fe489-config-volume\") pod \"collect-profiles-29483625-lsqb2\" (UID: \"c0b6ae2d-bfaf-4535-9906-31b19d4fe489\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.705225 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/bc0b3e27-43a2-4a5b-98bc-d371958b2bbf-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-q8wfq\" (UID: \"bc0b3e27-43a2-4a5b-98bc-d371958b2bbf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q8wfq" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.705273 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjphr\" (UniqueName: \"kubernetes.io/projected/b5ad8de7-53f2-475d-be7b-acb85ad7f53a-kube-api-access-gjphr\") pod \"multus-admission-controller-857f4d67dd-jxgs8\" (UID: \"b5ad8de7-53f2-475d-be7b-acb85ad7f53a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jxgs8" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.705288 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a23bae9-c21c-40a6-aa8d-9c3524614dd4-metrics-certs\") pod \"router-default-5444994796-xfblx\" (UID: \"9a23bae9-c21c-40a6-aa8d-9c3524614dd4\") " pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.708377 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/922aad90-2298-4afb-816e-063c73f18c6d-csi-data-dir\") pod \"csi-hostpathplugin-glhp5\" (UID: \"922aad90-2298-4afb-816e-063c73f18c6d\") " pod="hostpath-provisioner/csi-hostpathplugin-glhp5" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.709752 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/77ae37d7-ba3c-4b77-85cb-f7cb648ab795-tmpfs\") pod \"packageserver-d55dfcdfc-qfggh\" (UID: \"77ae37d7-ba3c-4b77-85cb-f7cb648ab795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.711600 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a23bae9-c21c-40a6-aa8d-9c3524614dd4-service-ca-bundle\") pod \"router-default-5444994796-xfblx\" (UID: \"9a23bae9-c21c-40a6-aa8d-9c3524614dd4\") " pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.712163 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/922aad90-2298-4afb-816e-063c73f18c6d-plugins-dir\") pod \"csi-hostpathplugin-glhp5\" (UID: \"922aad90-2298-4afb-816e-063c73f18c6d\") " pod="hostpath-provisioner/csi-hostpathplugin-glhp5" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.713301 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d98f2d52-51f0-4843-8ba3-50264b3b3abe-trusted-ca\") pod \"ingress-operator-5b745b69d9-mqm6k\" (UID: \"d98f2d52-51f0-4843-8ba3-50264b3b3abe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.721775 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c71840-39bc-4256-bdfb-43d6e1bb749e-config\") pod \"service-ca-operator-777779d784-fn8s9\" (UID: \"99c71840-39bc-4256-bdfb-43d6e1bb749e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fn8s9" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.721849 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/922aad90-2298-4afb-816e-063c73f18c6d-registration-dir\") pod \"csi-hostpathplugin-glhp5\" (UID: \"922aad90-2298-4afb-816e-063c73f18c6d\") " pod="hostpath-provisioner/csi-hostpathplugin-glhp5" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.723006 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0b6ae2d-bfaf-4535-9906-31b19d4fe489-secret-volume\") pod \"collect-profiles-29483625-lsqb2\" (UID: \"c0b6ae2d-bfaf-4535-9906-31b19d4fe489\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.724940 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/922aad90-2298-4afb-816e-063c73f18c6d-mountpoint-dir\") pod \"csi-hostpathplugin-glhp5\" (UID: \"922aad90-2298-4afb-816e-063c73f18c6d\") " pod="hostpath-provisioner/csi-hostpathplugin-glhp5" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.725224 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/044bc3f1-737c-4dda-9c24-d087ce039b42-config-volume\") pod \"dns-default-zjcb6\" (UID: \"044bc3f1-737c-4dda-9c24-d087ce039b42\") " pod="openshift-dns/dns-default-zjcb6" Jan 21 17:58:32 crc kubenswrapper[4808]: E0121 17:58:32.725716 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:33.225700587 +0000 UTC m=+147.956211542 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.725839 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c8a01e69-a842-4316-9fd6-a2850afa5b11-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-d2rkk\" (UID: \"c8a01e69-a842-4316-9fd6-a2850afa5b11\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d2rkk" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.731141 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/922aad90-2298-4afb-816e-063c73f18c6d-socket-dir\") pod \"csi-hostpathplugin-glhp5\" (UID: \"922aad90-2298-4afb-816e-063c73f18c6d\") " pod="hostpath-provisioner/csi-hostpathplugin-glhp5" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.731579 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e6e2b28f-7cfe-44a5-8a63-f17542f27cb6-signing-cabundle\") pod \"service-ca-9c57cc56f-s5khx\" (UID: \"e6e2b28f-7cfe-44a5-8a63-f17542f27cb6\") " pod="openshift-service-ca/service-ca-9c57cc56f-s5khx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.734968 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29185a07-7266-4975-870b-127f880f2cf6-cert\") pod \"ingress-canary-79g2n\" (UID: \"29185a07-7266-4975-870b-127f880f2cf6\") " pod="openshift-ingress-canary/ingress-canary-79g2n" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.738886 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0b6ae2d-bfaf-4535-9906-31b19d4fe489-config-volume\") pod \"collect-profiles-29483625-lsqb2\" (UID: \"c0b6ae2d-bfaf-4535-9906-31b19d4fe489\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.739193 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b5ad8de7-53f2-475d-be7b-acb85ad7f53a-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-jxgs8\" (UID: \"b5ad8de7-53f2-475d-be7b-acb85ad7f53a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jxgs8" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.739632 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-pvk7n\" (UID: \"5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5\") " pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.739815 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a23bae9-c21c-40a6-aa8d-9c3524614dd4-metrics-certs\") pod \"router-default-5444994796-xfblx\" (UID: \"9a23bae9-c21c-40a6-aa8d-9c3524614dd4\") " pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.740163 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9a23bae9-c21c-40a6-aa8d-9c3524614dd4-stats-auth\") pod \"router-default-5444994796-xfblx\" (UID: \"9a23bae9-c21c-40a6-aa8d-9c3524614dd4\") " pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.740701 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/77ae37d7-ba3c-4b77-85cb-f7cb648ab795-apiservice-cert\") pod \"packageserver-d55dfcdfc-qfggh\" (UID: \"77ae37d7-ba3c-4b77-85cb-f7cb648ab795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.740703 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wptxh\" (UniqueName: \"kubernetes.io/projected/2b3e892c-c86a-42cd-b257-d722cd033b5e-kube-api-access-wptxh\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.741390 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-pvk7n\" (UID: \"5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5\") " pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.741829 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d98f2d52-51f0-4843-8ba3-50264b3b3abe-metrics-tls\") pod \"ingress-operator-5b745b69d9-mqm6k\" (UID: \"d98f2d52-51f0-4843-8ba3-50264b3b3abe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.746951 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f669bbf5-0b37-4821-9b15-052eef33004e-node-bootstrap-token\") pod \"machine-config-server-2m5jx\" (UID: \"f669bbf5-0b37-4821-9b15-052eef33004e\") " pod="openshift-machine-config-operator/machine-config-server-2m5jx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.748310 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/77ae37d7-ba3c-4b77-85cb-f7cb648ab795-webhook-cert\") pod \"packageserver-d55dfcdfc-qfggh\" (UID: \"77ae37d7-ba3c-4b77-85cb-f7cb648ab795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.750032 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99c71840-39bc-4256-bdfb-43d6e1bb749e-serving-cert\") pod \"service-ca-operator-777779d784-fn8s9\" (UID: \"99c71840-39bc-4256-bdfb-43d6e1bb749e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fn8s9" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.750098 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c8a01e69-a842-4316-9fd6-a2850afa5b11-proxy-tls\") pod \"machine-config-controller-84d6567774-d2rkk\" (UID: \"c8a01e69-a842-4316-9fd6-a2850afa5b11\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d2rkk" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.750357 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/044bc3f1-737c-4dda-9c24-d087ce039b42-metrics-tls\") pod \"dns-default-zjcb6\" (UID: \"044bc3f1-737c-4dda-9c24-d087ce039b42\") " pod="openshift-dns/dns-default-zjcb6" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.750722 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9a23bae9-c21c-40a6-aa8d-9c3524614dd4-default-certificate\") pod \"router-default-5444994796-xfblx\" (UID: \"9a23bae9-c21c-40a6-aa8d-9c3524614dd4\") " pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.750891 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c6350524-fc0a-4bbc-b803-cea5fd8dc76c-srv-cert\") pod \"olm-operator-6b444d44fb-4gg7j\" (UID: \"c6350524-fc0a-4bbc-b803-cea5fd8dc76c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4gg7j" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.751728 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c6350524-fc0a-4bbc-b803-cea5fd8dc76c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4gg7j\" (UID: \"c6350524-fc0a-4bbc-b803-cea5fd8dc76c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4gg7j" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.752392 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e6e2b28f-7cfe-44a5-8a63-f17542f27cb6-signing-key\") pod \"service-ca-9c57cc56f-s5khx\" (UID: \"e6e2b28f-7cfe-44a5-8a63-f17542f27cb6\") " pod="openshift-service-ca/service-ca-9c57cc56f-s5khx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.779952 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/bc0b3e27-43a2-4a5b-98bc-d371958b2bbf-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-q8wfq\" (UID: \"bc0b3e27-43a2-4a5b-98bc-d371958b2bbf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q8wfq" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.803696 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d9f360c8-31a2-4577-8eec-d5d58c106933-metrics-tls\") pod \"dns-operator-744455d44c-z77hq\" (UID: \"d9f360c8-31a2-4577-8eec-d5d58c106933\") " pod="openshift-dns-operator/dns-operator-744455d44c-z77hq" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.804131 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/030a3f02-a9cd-4a3c-890d-686e2b7df274-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wj246\" (UID: \"030a3f02-a9cd-4a3c-890d-686e2b7df274\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wj246" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.804620 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f669bbf5-0b37-4821-9b15-052eef33004e-certs\") pod \"machine-config-server-2m5jx\" (UID: \"f669bbf5-0b37-4821-9b15-052eef33004e\") " pod="openshift-machine-config-operator/machine-config-server-2m5jx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.805021 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2c1a10ce-320b-48fe-9eca-ab5e6d110441-profile-collector-cert\") pod \"catalog-operator-68c6474976-jmp6c\" (UID: \"2c1a10ce-320b-48fe-9eca-ab5e6d110441\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jmp6c" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.805403 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2c1a10ce-320b-48fe-9eca-ab5e6d110441-srv-cert\") pod \"catalog-operator-68c6474976-jmp6c\" (UID: \"2c1a10ce-320b-48fe-9eca-ab5e6d110441\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jmp6c" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.805436 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmq7q\" (UniqueName: \"kubernetes.io/projected/d9f360c8-31a2-4577-8eec-d5d58c106933-kube-api-access-gmq7q\") pod \"dns-operator-744455d44c-z77hq\" (UID: \"d9f360c8-31a2-4577-8eec-d5d58c106933\") " pod="openshift-dns-operator/dns-operator-744455d44c-z77hq" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.805702 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqklb\" (UniqueName: \"kubernetes.io/projected/e897d5b0-11f6-4e3f-99d2-22a7d093738a-kube-api-access-hqklb\") pod \"machine-config-operator-74547568cd-n6tbv\" (UID: \"e897d5b0-11f6-4e3f-99d2-22a7d093738a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.806731 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:32 crc kubenswrapper[4808]: E0121 17:58:32.806817 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:33.306796505 +0000 UTC m=+148.037307390 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.807439 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b"] Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.808459 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: E0121 17:58:32.810801 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:33.310787249 +0000 UTC m=+148.041298144 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.812059 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hwlx\" (UniqueName: \"kubernetes.io/projected/e6e2b28f-7cfe-44a5-8a63-f17542f27cb6-kube-api-access-6hwlx\") pod \"service-ca-9c57cc56f-s5khx\" (UID: \"e6e2b28f-7cfe-44a5-8a63-f17542f27cb6\") " pod="openshift-service-ca/service-ca-9c57cc56f-s5khx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.832134 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hr5cr"] Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.839947 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8"] Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.846356 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7prvs\" (UniqueName: \"kubernetes.io/projected/5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5-kube-api-access-7prvs\") pod \"marketplace-operator-79b997595-pvk7n\" (UID: \"5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5\") " pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.861870 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d98f2d52-51f0-4843-8ba3-50264b3b3abe-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mqm6k\" (UID: \"d98f2d52-51f0-4843-8ba3-50264b3b3abe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.863505 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mr6rg\" (UniqueName: \"kubernetes.io/projected/c8a01e69-a842-4316-9fd6-a2850afa5b11-kube-api-access-mr6rg\") pod \"machine-config-controller-84d6567774-d2rkk\" (UID: \"c8a01e69-a842-4316-9fd6-a2850afa5b11\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d2rkk" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.886050 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqc87\" (UniqueName: \"kubernetes.io/projected/9a23bae9-c21c-40a6-aa8d-9c3524614dd4-kube-api-access-zqc87\") pod \"router-default-5444994796-xfblx\" (UID: \"9a23bae9-c21c-40a6-aa8d-9c3524614dd4\") " pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.901760 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp5kl\" (UniqueName: \"kubernetes.io/projected/bc0b3e27-43a2-4a5b-98bc-d371958b2bbf-kube-api-access-jp5kl\") pod \"package-server-manager-789f6589d5-q8wfq\" (UID: \"bc0b3e27-43a2-4a5b-98bc-d371958b2bbf\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q8wfq" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.913837 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:32 crc kubenswrapper[4808]: E0121 17:58:32.914129 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:33.414112225 +0000 UTC m=+148.144623110 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.914272 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:32 crc kubenswrapper[4808]: E0121 17:58:32.914881 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:33.414871836 +0000 UTC m=+148.145382721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.919499 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j99pq\" (UniqueName: \"kubernetes.io/projected/77ae37d7-ba3c-4b77-85cb-f7cb648ab795-kube-api-access-j99pq\") pod \"packageserver-d55dfcdfc-qfggh\" (UID: \"77ae37d7-ba3c-4b77-85cb-f7cb648ab795\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.936511 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-slth9" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.939201 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glmg8\" (UniqueName: \"kubernetes.io/projected/99c71840-39bc-4256-bdfb-43d6e1bb749e-kube-api-access-glmg8\") pod \"service-ca-operator-777779d784-fn8s9\" (UID: \"99c71840-39bc-4256-bdfb-43d6e1bb749e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fn8s9" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.941598 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.955832 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xcn7d" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.960070 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgd64\" (UniqueName: \"kubernetes.io/projected/c0b6ae2d-bfaf-4535-9906-31b19d4fe489-kube-api-access-bgd64\") pod \"collect-profiles-29483625-lsqb2\" (UID: \"c0b6ae2d-bfaf-4535-9906-31b19d4fe489\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.987524 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wj246" Jan 21 17:58:32 crc kubenswrapper[4808]: I0121 17:58:32.994916 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rdrb\" (UniqueName: \"kubernetes.io/projected/c6350524-fc0a-4bbc-b803-cea5fd8dc76c-kube-api-access-8rdrb\") pod \"olm-operator-6b444d44fb-4gg7j\" (UID: \"c6350524-fc0a-4bbc-b803-cea5fd8dc76c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4gg7j" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.013745 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-z77hq" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.014460 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fx7n\" (UniqueName: \"kubernetes.io/projected/922aad90-2298-4afb-816e-063c73f18c6d-kube-api-access-7fx7n\") pod \"csi-hostpathplugin-glhp5\" (UID: \"922aad90-2298-4afb-816e-063c73f18c6d\") " pod="hostpath-provisioner/csi-hostpathplugin-glhp5" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.018129 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:33 crc kubenswrapper[4808]: E0121 17:58:33.019294 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:33.519275367 +0000 UTC m=+148.249786252 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.027661 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.037386 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.056517 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q8wfq" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.057414 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bz45s"] Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.064815 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27k7s\" (UniqueName: \"kubernetes.io/projected/29185a07-7266-4975-870b-127f880f2cf6-kube-api-access-27k7s\") pod \"ingress-canary-79g2n\" (UID: \"29185a07-7266-4975-870b-127f880f2cf6\") " pod="openshift-ingress-canary/ingress-canary-79g2n" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.066026 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9fd6\" (UniqueName: \"kubernetes.io/projected/f669bbf5-0b37-4821-9b15-052eef33004e-kube-api-access-m9fd6\") pod \"machine-config-server-2m5jx\" (UID: \"f669bbf5-0b37-4821-9b15-052eef33004e\") " pod="openshift-machine-config-operator/machine-config-server-2m5jx" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.068474 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.077666 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4gg7j" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.084175 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5dpk\" (UniqueName: \"kubernetes.io/projected/044bc3f1-737c-4dda-9c24-d087ce039b42-kube-api-access-j5dpk\") pod \"dns-default-zjcb6\" (UID: \"044bc3f1-737c-4dda-9c24-d087ce039b42\") " pod="openshift-dns/dns-default-zjcb6" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.084530 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-s5khx" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.098630 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fn8s9" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.102897 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjphr\" (UniqueName: \"kubernetes.io/projected/b5ad8de7-53f2-475d-be7b-acb85ad7f53a-kube-api-access-gjphr\") pod \"multus-admission-controller-857f4d67dd-jxgs8\" (UID: \"b5ad8de7-53f2-475d-be7b-acb85ad7f53a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jxgs8" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.103114 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d2rkk" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.108977 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.112418 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh256\" (UniqueName: \"kubernetes.io/projected/2c1a10ce-320b-48fe-9eca-ab5e6d110441-kube-api-access-nh256\") pod \"catalog-operator-68c6474976-jmp6c\" (UID: \"2c1a10ce-320b-48fe-9eca-ab5e6d110441\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jmp6c" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.115715 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4g7bd"] Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.120918 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:33 crc kubenswrapper[4808]: E0121 17:58:33.121280 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:33.621267177 +0000 UTC m=+148.351778052 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.125358 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-ftpvl"] Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.127506 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-glhp5" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.140566 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jcnqg"] Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.149607 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2m5jx" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.150004 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m679v\" (UniqueName: \"kubernetes.io/projected/d98f2d52-51f0-4843-8ba3-50264b3b3abe-kube-api-access-m679v\") pod \"ingress-operator-5b745b69d9-mqm6k\" (UID: \"d98f2d52-51f0-4843-8ba3-50264b3b3abe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.152991 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-zjcb6" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.153560 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-79g2n" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.227863 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:33 crc kubenswrapper[4808]: E0121 17:58:33.228686 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:33.72866479 +0000 UTC m=+148.459175675 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.294399 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" podStartSLOduration=125.294382513 podStartE2EDuration="2m5.294382513s" podCreationTimestamp="2026-01-21 17:56:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:33.236742453 +0000 UTC m=+147.967253338" watchObservedRunningTime="2026-01-21 17:58:33.294382513 +0000 UTC m=+148.024893638" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.305467 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-jxgs8" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.322684 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.329320 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:33 crc kubenswrapper[4808]: E0121 17:58:33.339637 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:33.839228374 +0000 UTC m=+148.569739259 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.348117 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jmp6c" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.364747 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lq8cg"] Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.381536 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4g7bd" event={"ID":"4c87e4f3-6255-4fe3-a52c-726a1acf7ba8","Type":"ContainerStarted","Data":"fc771de52a63dd19c2746c9acd77c8b14bfeb261e45fd0cfd59b3f006cb0fc01"} Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.408560 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"30acb34b2f95d5563b6cca7d00a5631728c17a94783d2783cee1d7cde382b8d6"} Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.409952 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-jcnqg" event={"ID":"05d34d86-bf40-4b28-94b9-8292fe685637","Type":"ContainerStarted","Data":"c97e8e7e3d60de4c600b7075da418758a1736810e63e2f59f15c3385931f55df"} Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.421083 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hr5cr" event={"ID":"26022c12-cb37-40ed-9ecb-aef5f82d2bed","Type":"ContainerStarted","Data":"e686b57c11d8dc58e6e503389fee7ed8c806de8016241ec5a5e19478380afce3"} Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.421539 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-q95v9"] Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.432164 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:33 crc kubenswrapper[4808]: E0121 17:58:33.432601 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:33.932584038 +0000 UTC m=+148.663094923 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.454750 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b" event={"ID":"1c1e9bbd-24f9-4ef3-8844-6ddb9861c509","Type":"ContainerStarted","Data":"bdddbde03784b24aeb48f6fcced8c22ba9dad56189bff75690b6fa926777c975"} Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.454811 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b" event={"ID":"1c1e9bbd-24f9-4ef3-8844-6ddb9861c509","Type":"ContainerStarted","Data":"700a604ffd6056bcae34fd8a27ea3ebab4471ae94399b41fd23d7b1a530756b3"} Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.456256 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" event={"ID":"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5","Type":"ContainerStarted","Data":"7679e7a24ca34b070746c965775003ef9ae5bd5ca74a3c7385ded5a88b316e30"} Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.474003 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"726d2fae2a9d130c79ae14684ce6096b47e2159eafc9e764e72e2bb36e0ac95c"} Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.474364 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.476269 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ftpvl" event={"ID":"196e1220-375f-4a05-90e0-ce571f606de4","Type":"ContainerStarted","Data":"b6165dae517faa7368feff5812408156c390b9bafcafb733f6afefd997d84799"} Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.477177 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" event={"ID":"66c5c3c8-4515-401a-bfe9-5755068bc771","Type":"ContainerStarted","Data":"af2baea7f35d573815130e53461594fc2a9014abe94f43e837d136de3413e5a8"} Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.478287 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-fc6zr" event={"ID":"a189dfc0-4fe2-4290-b881-f238a055ab5a","Type":"ContainerStarted","Data":"7f58294aaa2384865c13d56cd165a72f61c9386547cfc53825adceea1fa8b412"} Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.478317 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-fc6zr" event={"ID":"a189dfc0-4fe2-4290-b881-f238a055ab5a","Type":"ContainerStarted","Data":"71ad5255bd0e30c853821b7ae05a3d670a377bd08bf7e98a7a5102f56fd9a249"} Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.480691 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-fc6zr" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.484159 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ps7kv" event={"ID":"ab1aa451-ae2a-48ec-ae13-0689bad3c894","Type":"ContainerStarted","Data":"846611b8da7dc0b5d80ebe14dac11d0f003a529aed32dca1b30b2027ef4a9f6e"} Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.484216 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ps7kv" event={"ID":"ab1aa451-ae2a-48ec-ae13-0689bad3c894","Type":"ContainerStarted","Data":"0027e2003665b08654e4558f20cd4b7a84f088f64535411b112e64601eccfa4e"} Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.486008 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" event={"ID":"3919133e-aaba-4dc1-b4cb-6b09f6440221","Type":"ContainerStarted","Data":"4c6a75571185d2caa1a47120b61b4c6781115c979189cb3b908243d3d19c82b5"} Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.486038 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" event={"ID":"3919133e-aaba-4dc1-b4cb-6b09f6440221","Type":"ContainerStarted","Data":"0fafb665cfc5ffb6789f9e9c4f7e96b7efe3a1a1c6c27490504d9a8d24447553"} Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.488945 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.505236 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-h6gm7" event={"ID":"05688b08-2c9f-4c6c-967a-ff64de7837c0","Type":"ContainerStarted","Data":"1a742fbf455b96cc61fe501e3c5e2c27d7233fd192beace247a544eef0f3efc6"} Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.505293 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-h6gm7" event={"ID":"05688b08-2c9f-4c6c-967a-ff64de7837c0","Type":"ContainerStarted","Data":"ee872900240b24a7b79e371b192d04a17145bdeed574bfe0d1dcbf718ddd32b0"} Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.510587 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.537329 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:33 crc kubenswrapper[4808]: E0121 17:58:33.539948 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:34.039908099 +0000 UTC m=+148.770418984 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.641910 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:33 crc kubenswrapper[4808]: E0121 17:58:33.643479 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:34.143459374 +0000 UTC m=+148.873970269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.749138 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:33 crc kubenswrapper[4808]: E0121 17:58:33.749735 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:34.24971941 +0000 UTC m=+148.980230295 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.853744 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:33 crc kubenswrapper[4808]: E0121 17:58:33.854264 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:34.354226025 +0000 UTC m=+149.084736910 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.877207 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-628ks"] Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.883059 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-fc6zr" Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.954975 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:33 crc kubenswrapper[4808]: E0121 17:58:33.955380 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:34.455365 +0000 UTC m=+149.185875905 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:33 crc kubenswrapper[4808]: I0121 17:58:33.983545 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-blj6k"] Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.056602 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:34 crc kubenswrapper[4808]: E0121 17:58:34.056961 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:34.556945483 +0000 UTC m=+149.287456358 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.104184 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" podStartSLOduration=125.104163463 podStartE2EDuration="2m5.104163463s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:34.102671261 +0000 UTC m=+148.833182156" watchObservedRunningTime="2026-01-21 17:58:34.104163463 +0000 UTC m=+148.834674348" Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.162330 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:34 crc kubenswrapper[4808]: E0121 17:58:34.162976 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:34.662943599 +0000 UTC m=+149.393454484 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.277387 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:34 crc kubenswrapper[4808]: E0121 17:58:34.278442 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:34.778414116 +0000 UTC m=+149.508925001 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.335588 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cn7qm"] Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.351545 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jl78z"] Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.363780 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8jqv"] Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.383437 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:34 crc kubenswrapper[4808]: E0121 17:58:34.384091 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:34.884076688 +0000 UTC m=+149.614587573 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.488693 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:34 crc kubenswrapper[4808]: E0121 17:58:34.489148 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:34.989130785 +0000 UTC m=+149.719641670 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:34 crc kubenswrapper[4808]: W0121 17:58:34.510742 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11782075_8384_415d_a1b9_6c972f8f6f66.slice/crio-eeac24d2d8146070d5ce61f53751f3a3657060978587107c302781c4e9ecb6e7 WatchSource:0}: Error finding container eeac24d2d8146070d5ce61f53751f3a3657060978587107c302781c4e9ecb6e7: Status 404 returned error can't find the container with id eeac24d2d8146070d5ce61f53751f3a3657060978587107c302781c4e9ecb6e7 Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.589865 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.590205 4808 generic.go:334] "Generic (PLEG): container finished" podID="1c1e9bbd-24f9-4ef3-8844-6ddb9861c509" containerID="bdddbde03784b24aeb48f6fcced8c22ba9dad56189bff75690b6fa926777c975" exitCode=0 Jan 21 17:58:34 crc kubenswrapper[4808]: E0121 17:58:34.590946 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:35.090930968 +0000 UTC m=+149.821441853 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.591000 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b" event={"ID":"1c1e9bbd-24f9-4ef3-8844-6ddb9861c509","Type":"ContainerDied","Data":"bdddbde03784b24aeb48f6fcced8c22ba9dad56189bff75690b6fa926777c975"} Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.596172 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b" event={"ID":"1c1e9bbd-24f9-4ef3-8844-6ddb9861c509","Type":"ContainerStarted","Data":"e4bd522f4177ce80c2eb1def2e6e8aa332d37140436a7a717e12f33afb1a88bf"} Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.596264 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b" Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.606386 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ftpvl" event={"ID":"196e1220-375f-4a05-90e0-ce571f606de4","Type":"ContainerStarted","Data":"c2079c4974dccfaab9f395fe5ce5e48dae1c45051abe8c18ecb9ac1866f81f6a"} Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.608779 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-ftpvl" Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.630682 4808 patch_prober.go:28] interesting pod/downloads-7954f5f757-ftpvl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.630754 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ftpvl" podUID="196e1220-375f-4a05-90e0-ce571f606de4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.632712 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-blj6k" event={"ID":"afb48788-1b96-4ba0-9a9d-98e613451c6f","Type":"ContainerStarted","Data":"dc586809c69f1b385afd0655af96b894a5c6040c7bcffb408da07e5068dbae5c"} Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.643731 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8jqv" event={"ID":"871099ce-cd86-498e-b227-ec40b6c29e47","Type":"ContainerStarted","Data":"ec2fc29f04a11764929de585bd21bad71a5d77a7ccee9cc154bb6c1160d1c6a7"} Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.660074 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" event={"ID":"56d0bac1-fb03-49ab-8e45-1cb28643dc04","Type":"ContainerStarted","Data":"1059dfe8cf530b73969b3f6ccdc69d1bc82379614aecdbe5740bac250c5285fa"} Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.680438 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lq8cg" event={"ID":"5b91a9d5-e508-4624-a43c-770a9dcdbc1d","Type":"ContainerStarted","Data":"5d1e45d56d8e5f4d79bcaf5bea9d613b4e8d572ed9306e292de6985308c06e12"} Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.695309 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jl78z" event={"ID":"11782075-8384-415d-a1b9-6c972f8f6f66","Type":"ContainerStarted","Data":"eeac24d2d8146070d5ce61f53751f3a3657060978587107c302781c4e9ecb6e7"} Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.697558 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:34 crc kubenswrapper[4808]: E0121 17:58:34.698578 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:35.198562151 +0000 UTC m=+149.929073026 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.733724 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xcn7d"] Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.745282 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fn8s9"] Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.769603 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2m5jx" event={"ID":"f669bbf5-0b37-4821-9b15-052eef33004e","Type":"ContainerStarted","Data":"902e54bb86f5970acf44009217cc97a6a474a02ef314cb34fce3ccc97a8983a3"} Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.769654 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2m5jx" event={"ID":"f669bbf5-0b37-4821-9b15-052eef33004e","Type":"ContainerStarted","Data":"8872f4b5733538484ebc9cefd16afb7268331d379c078bf8f67fbad33053e175"} Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.785523 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b" podStartSLOduration=125.78550182 podStartE2EDuration="2m5.78550182s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:34.782788088 +0000 UTC m=+149.513298973" watchObservedRunningTime="2026-01-21 17:58:34.78550182 +0000 UTC m=+149.516012705" Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.793502 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4g7bd" event={"ID":"4c87e4f3-6255-4fe3-a52c-726a1acf7ba8","Type":"ContainerStarted","Data":"1baaaf98f4bf54dd431794e03fa97312ee9b3458e3ac7d44e4626dbb8d899d04"} Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.799062 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:34 crc kubenswrapper[4808]: E0121 17:58:34.800189 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:35.300177046 +0000 UTC m=+150.030687931 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.812988 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ps7kv" podStartSLOduration=126.812967624 podStartE2EDuration="2m6.812967624s" podCreationTimestamp="2026-01-21 17:56:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:34.806705165 +0000 UTC m=+149.537216050" watchObservedRunningTime="2026-01-21 17:58:34.812967624 +0000 UTC m=+149.543478509" Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.818236 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-xfblx" event={"ID":"9a23bae9-c21c-40a6-aa8d-9c3524614dd4","Type":"ContainerStarted","Data":"8ed2bb6e52f180fc6e1a531d0efa3de3fbc5d115e9ef67c2ef9edf20f48c25f5"} Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.818300 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-xfblx" event={"ID":"9a23bae9-c21c-40a6-aa8d-9c3524614dd4","Type":"ContainerStarted","Data":"56025cfdfa7cbc4326a003bbdd6916e97be76aaf66ea45d479bbb1d69fa952de"} Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.828944 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hr5cr" event={"ID":"26022c12-cb37-40ed-9ecb-aef5f82d2bed","Type":"ContainerStarted","Data":"c47dd539f90f47c844897121a9f295e1cc5b98454eb284becbbcb54a4e74c45e"} Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.836763 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.840020 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" event={"ID":"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5","Type":"ContainerStarted","Data":"c0cabeb627bf3a86c247f34d3e1a6343af8dcaadbc9bf25abef9963c2a42a8e4"} Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.841595 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.848176 4808 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-bz45s container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.16:6443/healthz\": dial tcp 10.217.0.16:6443: connect: connection refused" start-of-body= Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.848264 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" podUID="66c5c3c8-4515-401a-bfe9-5755068bc771" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.16:6443/healthz\": dial tcp 10.217.0.16:6443: connect: connection refused" Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.852270 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q95v9" event={"ID":"dbc348aa-a6f2-4fee-8c96-83f98c8358b9","Type":"ContainerStarted","Data":"64a015bcbcbeca7e39339fa06d356532016fa6f5552139867a02d3ea6b84b3b9"} Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.852378 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q95v9" event={"ID":"dbc348aa-a6f2-4fee-8c96-83f98c8358b9","Type":"ContainerStarted","Data":"1e652e5e8995bd5d6753fc43065a178a711401894d02fc55ad763676c124325c"} Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.857804 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.868100 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-jcnqg" event={"ID":"05d34d86-bf40-4b28-94b9-8292fe685637","Type":"ContainerStarted","Data":"00aac258b6d0317684c99f27180158189a5e44205b1029a37163ec39f1a99abc"} Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.876160 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv"] Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.900892 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-slth9"] Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.901002 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:34 crc kubenswrapper[4808]: E0121 17:58:34.902628 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:35.402609945 +0000 UTC m=+150.133120830 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.944023 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" podStartSLOduration=125.920217592 podStartE2EDuration="2m5.920217592s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:34.917924077 +0000 UTC m=+149.648434982" watchObservedRunningTime="2026-01-21 17:58:34.920217592 +0000 UTC m=+149.650728477" Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.956871 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-fc6zr" podStartSLOduration=125.956843354 podStartE2EDuration="2m5.956843354s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:34.950651068 +0000 UTC m=+149.681161953" watchObservedRunningTime="2026-01-21 17:58:34.956843354 +0000 UTC m=+149.687354239" Jan 21 17:58:34 crc kubenswrapper[4808]: I0121 17:58:34.985700 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-ftpvl" podStartSLOduration=125.985683404 podStartE2EDuration="2m5.985683404s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:34.985451285 +0000 UTC m=+149.715962160" watchObservedRunningTime="2026-01-21 17:58:34.985683404 +0000 UTC m=+149.716194289" Jan 21 17:58:35 crc kubenswrapper[4808]: E0121 17:58:35.011138 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:35.511124224 +0000 UTC m=+150.241635109 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.019187 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.030171 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.038772 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-h6gm7" podStartSLOduration=126.038750425 podStartE2EDuration="2m6.038750425s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:35.038278605 +0000 UTC m=+149.768789500" watchObservedRunningTime="2026-01-21 17:58:35.038750425 +0000 UTC m=+149.769261310" Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.045426 4808 patch_prober.go:28] interesting pod/router-default-5444994796-xfblx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 17:58:35 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Jan 21 17:58:35 crc kubenswrapper[4808]: [+]process-running ok Jan 21 17:58:35 crc kubenswrapper[4808]: healthz check failed Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.045754 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xfblx" podUID="9a23bae9-c21c-40a6-aa8d-9c3524614dd4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.101024 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-4g7bd" podStartSLOduration=126.100998945 podStartE2EDuration="2m6.100998945s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:35.075519613 +0000 UTC m=+149.806030488" watchObservedRunningTime="2026-01-21 17:58:35.100998945 +0000 UTC m=+149.831509840" Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.104196 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wj246"] Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.129371 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:35 crc kubenswrapper[4808]: E0121 17:58:35.130819 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:35.630794865 +0000 UTC m=+150.361305750 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.164640 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" podStartSLOduration=126.164623451 podStartE2EDuration="2m6.164623451s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:35.113637096 +0000 UTC m=+149.844147991" watchObservedRunningTime="2026-01-21 17:58:35.164623451 +0000 UTC m=+149.895134336" Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.165156 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" podStartSLOduration=127.165151183 podStartE2EDuration="2m7.165151183s" podCreationTimestamp="2026-01-21 17:56:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:35.155702163 +0000 UTC m=+149.886213058" watchObservedRunningTime="2026-01-21 17:58:35.165151183 +0000 UTC m=+149.895662058" Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.198008 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z77hq"] Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.200566 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-jcnqg" podStartSLOduration=127.200540644 podStartE2EDuration="2m7.200540644s" podCreationTimestamp="2026-01-21 17:56:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:35.194117679 +0000 UTC m=+149.924628564" watchObservedRunningTime="2026-01-21 17:58:35.200540644 +0000 UTC m=+149.931051529" Jan 21 17:58:35 crc kubenswrapper[4808]: W0121 17:58:35.224166 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod030a3f02_a9cd_4a3c_890d_686e2b7df274.slice/crio-f525a04fd11167455266f22edf3ffb81999161b66ccb5d23aaffa95a13230f1e WatchSource:0}: Error finding container f525a04fd11167455266f22edf3ffb81999161b66ccb5d23aaffa95a13230f1e: Status 404 returned error can't find the container with id f525a04fd11167455266f22edf3ffb81999161b66ccb5d23aaffa95a13230f1e Jan 21 17:58:35 crc kubenswrapper[4808]: W0121 17:58:35.224413 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9f360c8_31a2_4577_8eec_d5d58c106933.slice/crio-6a164dd8ebb98df4ae956160f264ba6766b503c7b1309888abd54e59dc9516b6 WatchSource:0}: Error finding container 6a164dd8ebb98df4ae956160f264ba6766b503c7b1309888abd54e59dc9516b6: Status 404 returned error can't find the container with id 6a164dd8ebb98df4ae956160f264ba6766b503c7b1309888abd54e59dc9516b6 Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.237743 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-2m5jx" podStartSLOduration=5.237728749 podStartE2EDuration="5.237728749s" podCreationTimestamp="2026-01-21 17:58:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:35.237231109 +0000 UTC m=+149.967741994" watchObservedRunningTime="2026-01-21 17:58:35.237728749 +0000 UTC m=+149.968239634" Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.238294 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:35 crc kubenswrapper[4808]: E0121 17:58:35.238593 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:35.738581364 +0000 UTC m=+150.469092249 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.262509 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.271561 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.272768 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hr5cr" podStartSLOduration=127.272227754 podStartE2EDuration="2m7.272227754s" podCreationTimestamp="2026-01-21 17:56:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:35.271063915 +0000 UTC m=+150.001574810" watchObservedRunningTime="2026-01-21 17:58:35.272227754 +0000 UTC m=+150.002738639" Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.307537 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q8wfq"] Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.309702 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-xfblx" podStartSLOduration=126.309674899 podStartE2EDuration="2m6.309674899s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:35.308598505 +0000 UTC m=+150.039109390" watchObservedRunningTime="2026-01-21 17:58:35.309674899 +0000 UTC m=+150.040185774" Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.327386 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.340774 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:35 crc kubenswrapper[4808]: E0121 17:58:35.341166 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:35.841150989 +0000 UTC m=+150.571661874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.355536 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-jxgs8"] Jan 21 17:58:35 crc kubenswrapper[4808]: W0121 17:58:35.376120 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc0b3e27_43a2_4a5b_98bc_d371958b2bbf.slice/crio-80abd290f45929d8a82977c401c3af47d2a2a253d798f2fffe8fef6511eda830 WatchSource:0}: Error finding container 80abd290f45929d8a82977c401c3af47d2a2a253d798f2fffe8fef6511eda830: Status 404 returned error can't find the container with id 80abd290f45929d8a82977c401c3af47d2a2a253d798f2fffe8fef6511eda830 Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.393956 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k"] Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.414615 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh"] Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.442007 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-d2rkk"] Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.442849 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:35 crc kubenswrapper[4808]: E0121 17:58:35.445089 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:35.945074609 +0000 UTC m=+150.675585484 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.479104 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-zjcb6"] Jan 21 17:58:35 crc kubenswrapper[4808]: W0121 17:58:35.521405 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8a01e69_a842_4316_9fd6_a2850afa5b11.slice/crio-69ba3afbe5a180dd3fec2230611ceb2f10eb6e1009e95ae9e9a7e09e17c448a6 WatchSource:0}: Error finding container 69ba3afbe5a180dd3fec2230611ceb2f10eb6e1009e95ae9e9a7e09e17c448a6: Status 404 returned error can't find the container with id 69ba3afbe5a180dd3fec2230611ceb2f10eb6e1009e95ae9e9a7e09e17c448a6 Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.556517 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:35 crc kubenswrapper[4808]: E0121 17:58:35.556866 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:36.056852194 +0000 UTC m=+150.787363079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.593414 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2"] Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.593466 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.593485 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.593499 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-glhp5"] Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.593514 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-79g2n"] Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.593527 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pvk7n"] Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.594006 4808 csr.go:261] certificate signing request csr-z87bm is approved, waiting to be issued Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.596367 4808 csr.go:257] certificate signing request csr-z87bm is issued Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.633725 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.661333 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:35 crc kubenswrapper[4808]: E0121 17:58:35.664333 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:36.16431818 +0000 UTC m=+150.894829065 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.689909 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-s5khx"] Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.689969 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4gg7j"] Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.701732 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jmp6c"] Jan 21 17:58:35 crc kubenswrapper[4808]: W0121 17:58:35.751340 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0b6ae2d_bfaf_4535_9906_31b19d4fe489.slice/crio-a265a7647c478ac03358f72e533f89555b59cf46ff052c2a97d976b34c3410bd WatchSource:0}: Error finding container a265a7647c478ac03358f72e533f89555b59cf46ff052c2a97d976b34c3410bd: Status 404 returned error can't find the container with id a265a7647c478ac03358f72e533f89555b59cf46ff052c2a97d976b34c3410bd Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.767956 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:35 crc kubenswrapper[4808]: E0121 17:58:35.768549 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:36.268522142 +0000 UTC m=+150.999033027 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.768680 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:35 crc kubenswrapper[4808]: E0121 17:58:35.769030 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:36.269015762 +0000 UTC m=+150.999526647 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:35 crc kubenswrapper[4808]: W0121 17:58:35.794382 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c1a10ce_320b_48fe_9eca_ab5e6d110441.slice/crio-fd2d3c2373b6c21dd472c1e969334ea9fb526ff8855c310eea2035a6b2ee3030 WatchSource:0}: Error finding container fd2d3c2373b6c21dd472c1e969334ea9fb526ff8855c310eea2035a6b2ee3030: Status 404 returned error can't find the container with id fd2d3c2373b6c21dd472c1e969334ea9fb526ff8855c310eea2035a6b2ee3030 Jan 21 17:58:35 crc kubenswrapper[4808]: W0121 17:58:35.804330 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6e2b28f_7cfe_44a5_8a63_f17542f27cb6.slice/crio-67f877fede91f3aaec0fca716927cf06d43fbb71f60b680524d5aadbdf54d5f2 WatchSource:0}: Error finding container 67f877fede91f3aaec0fca716927cf06d43fbb71f60b680524d5aadbdf54d5f2: Status 404 returned error can't find the container with id 67f877fede91f3aaec0fca716927cf06d43fbb71f60b680524d5aadbdf54d5f2 Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.869343 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:35 crc kubenswrapper[4808]: E0121 17:58:35.869653 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:36.369636016 +0000 UTC m=+151.100146901 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.897389 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q8wfq" event={"ID":"bc0b3e27-43a2-4a5b-98bc-d371958b2bbf","Type":"ContainerStarted","Data":"80abd290f45929d8a82977c401c3af47d2a2a253d798f2fffe8fef6511eda830"} Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.903120 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv" event={"ID":"e897d5b0-11f6-4e3f-99d2-22a7d093738a","Type":"ContainerStarted","Data":"5b5cfcfec2ca8e611bf87966fa616b7d2e8a37310cae824b026c6a63989dc37b"} Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.903220 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv" event={"ID":"e897d5b0-11f6-4e3f-99d2-22a7d093738a","Type":"ContainerStarted","Data":"456584dd15519796ebb036626d214cd48ed6d38b9d31cff45c5e65122dfb8abf"} Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.907771 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8jqv" event={"ID":"871099ce-cd86-498e-b227-ec40b6c29e47","Type":"ContainerStarted","Data":"00a880ba617ad384345e3dc1b6b38be44ad78a7bd540802cfc549e7a14a64f62"} Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.924021 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" event={"ID":"66c5c3c8-4515-401a-bfe9-5755068bc771","Type":"ContainerStarted","Data":"c566ddd31fa616ed33feec6e779ffa61917579a6a973416a023738d2e73a78a4"} Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.930364 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh" event={"ID":"77ae37d7-ba3c-4b77-85cb-f7cb648ab795","Type":"ContainerStarted","Data":"d23b453bcdc79b92a538e79b4345e1fd6d39326b28f2f2d7a282d5d619f9e8b0"} Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.948625 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.949477 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-slth9" event={"ID":"433911de-428a-4261-8680-66a30c308919","Type":"ContainerStarted","Data":"27b528f9918c698b9342337d46be7a081ab995a5bb19eb6edfa840ee6551fb0f"} Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.971720 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:35 crc kubenswrapper[4808]: E0121 17:58:35.972126 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:36.472114927 +0000 UTC m=+151.202625812 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.976538 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fn8s9" event={"ID":"99c71840-39bc-4256-bdfb-43d6e1bb749e","Type":"ContainerStarted","Data":"db05f34d6d50b0244a4d651702e9174825abf436ba61e83cc9a0cbee6cabe1fe"} Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.976600 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fn8s9" event={"ID":"99c71840-39bc-4256-bdfb-43d6e1bb749e","Type":"ContainerStarted","Data":"3f266ae26528e4cada5b728fc128c66dff5fb20f2f757bc39dcb741d816eed49"} Jan 21 17:58:35 crc kubenswrapper[4808]: I0121 17:58:35.989822 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k" event={"ID":"d98f2d52-51f0-4843-8ba3-50264b3b3abe","Type":"ContainerStarted","Data":"e0740b02b72feb01f90da0a7b8e433b27e496818730c88ea716a85452bb5f16d"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.002142 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z77hq" event={"ID":"d9f360c8-31a2-4577-8eec-d5d58c106933","Type":"ContainerStarted","Data":"6a164dd8ebb98df4ae956160f264ba6766b503c7b1309888abd54e59dc9516b6"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.005400 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wj246" event={"ID":"030a3f02-a9cd-4a3c-890d-686e2b7df274","Type":"ContainerStarted","Data":"f525a04fd11167455266f22edf3ffb81999161b66ccb5d23aaffa95a13230f1e"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.070828 4808 patch_prober.go:28] interesting pod/router-default-5444994796-xfblx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 17:58:36 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Jan 21 17:58:36 crc kubenswrapper[4808]: [+]process-running ok Jan 21 17:58:36 crc kubenswrapper[4808]: healthz check failed Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.070931 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xfblx" podUID="9a23bae9-c21c-40a6-aa8d-9c3524614dd4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.074237 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:36 crc kubenswrapper[4808]: E0121 17:58:36.075614 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:36.575589779 +0000 UTC m=+151.306100664 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.082799 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:36 crc kubenswrapper[4808]: E0121 17:58:36.085938 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:36.585915265 +0000 UTC m=+151.316426150 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.089818 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q95v9" event={"ID":"dbc348aa-a6f2-4fee-8c96-83f98c8358b9","Type":"ContainerStarted","Data":"506794bc22b2fd7251237b556730e86d3355fa885cea1b4d622ef0cd2d673a49"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.154845 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q95v9" podStartSLOduration=127.15482689 podStartE2EDuration="2m7.15482689s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:36.154473015 +0000 UTC m=+150.884983900" watchObservedRunningTime="2026-01-21 17:58:36.15482689 +0000 UTC m=+150.885337775" Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.180284 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" event={"ID":"56d0bac1-fb03-49ab-8e45-1cb28643dc04","Type":"ContainerStarted","Data":"861c78b7691effc52be4c408732e9f82228f55df15d086e53defcc2fbb46bbcd"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.183850 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:36 crc kubenswrapper[4808]: E0121 17:58:36.184128 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:36.684106448 +0000 UTC m=+151.414617343 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.184264 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:36 crc kubenswrapper[4808]: E0121 17:58:36.185354 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:36.685343379 +0000 UTC m=+151.415854324 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.197501 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2" event={"ID":"c0b6ae2d-bfaf-4535-9906-31b19d4fe489","Type":"ContainerStarted","Data":"a265a7647c478ac03358f72e533f89555b59cf46ff052c2a97d976b34c3410bd"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.218132 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4g7bd" event={"ID":"4c87e4f3-6255-4fe3-a52c-726a1acf7ba8","Type":"ContainerStarted","Data":"5126fdef97251e093ef80d101ada110b0324c9dc838da27016de79cbae18f669"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.253939 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fn8s9" podStartSLOduration=127.25392025 podStartE2EDuration="2m7.25392025s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:36.19771768 +0000 UTC m=+150.928228565" watchObservedRunningTime="2026-01-21 17:58:36.25392025 +0000 UTC m=+150.984431135" Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.278993 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-jxgs8" event={"ID":"b5ad8de7-53f2-475d-be7b-acb85ad7f53a","Type":"ContainerStarted","Data":"bb61bd5cac2fa45d24e933c4854f4f0f0c8d73e5b8a43514d23f458aa3bc59d7"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.285796 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:36 crc kubenswrapper[4808]: E0121 17:58:36.286491 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:36.786470244 +0000 UTC m=+151.516981129 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.294269 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-79g2n" event={"ID":"29185a07-7266-4975-870b-127f880f2cf6","Type":"ContainerStarted","Data":"8926a2bcdaef6c01a54658c44f807472d6d8348148cf9d94a3283a4fd17bcd80"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.331909 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d2rkk" event={"ID":"c8a01e69-a842-4316-9fd6-a2850afa5b11","Type":"ContainerStarted","Data":"69ba3afbe5a180dd3fec2230611ceb2f10eb6e1009e95ae9e9a7e09e17c448a6"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.336023 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jmp6c" event={"ID":"2c1a10ce-320b-48fe-9eca-ab5e6d110441","Type":"ContainerStarted","Data":"fd2d3c2373b6c21dd472c1e969334ea9fb526ff8855c310eea2035a6b2ee3030"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.365723 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8jqv" podStartSLOduration=127.365703555 podStartE2EDuration="2m7.365703555s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:36.363719813 +0000 UTC m=+151.094230698" watchObservedRunningTime="2026-01-21 17:58:36.365703555 +0000 UTC m=+151.096214440" Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.366006 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jl78z" event={"ID":"11782075-8384-415d-a1b9-6c972f8f6f66","Type":"ContainerStarted","Data":"42ba956661ca88d39077a81f535292a3d33688ad0e698c903b0d0cdc880929cf"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.382944 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4gg7j" event={"ID":"c6350524-fc0a-4bbc-b803-cea5fd8dc76c","Type":"ContainerStarted","Data":"14e578bc57fe0e6a782aee89298e1e8e95f2c63cf17659dac0ebbcd4ddef0d64"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.387689 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:36 crc kubenswrapper[4808]: E0121 17:58:36.389092 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:36.8890767 +0000 UTC m=+151.619587585 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.408746 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-glhp5" event={"ID":"922aad90-2298-4afb-816e-063c73f18c6d","Type":"ContainerStarted","Data":"3cf08f4eadbfc3fd249488bee38babb9ce129d66de1438c0d3242dd00fbd9f73"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.411837 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-slth9" podStartSLOduration=127.411818589 podStartE2EDuration="2m7.411818589s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:36.40941485 +0000 UTC m=+151.139925745" watchObservedRunningTime="2026-01-21 17:58:36.411818589 +0000 UTC m=+151.142329474" Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.413726 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xcn7d" event={"ID":"a848b67a-a4cd-4057-8a2e-2587d6b7142d","Type":"ContainerStarted","Data":"c75290dfc22bca511263105973b64ca59db15a7ddf100d08692f4f0114b7e9c9"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.413783 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xcn7d" event={"ID":"a848b67a-a4cd-4057-8a2e-2587d6b7142d","Type":"ContainerStarted","Data":"a3f35c12c6a55306134b9e4d2078f02a7f0942a66ea2e26523c79429f4452781"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.443849 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-628ks" podStartSLOduration=127.44383023 podStartE2EDuration="2m7.44383023s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:36.442651952 +0000 UTC m=+151.173162847" watchObservedRunningTime="2026-01-21 17:58:36.44383023 +0000 UTC m=+151.174341115" Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.459539 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lq8cg" event={"ID":"5b91a9d5-e508-4624-a43c-770a9dcdbc1d","Type":"ContainerStarted","Data":"ef9714034e0512dc8bd4fd86164b3e1d9273761684df27204df157b7db2697a3"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.488677 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:36 crc kubenswrapper[4808]: E0121 17:58:36.488957 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:36.988931772 +0000 UTC m=+151.719442657 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.489173 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:36 crc kubenswrapper[4808]: E0121 17:58:36.489579 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:36.989563419 +0000 UTC m=+151.720074304 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.528666 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-blj6k" event={"ID":"afb48788-1b96-4ba0-9a9d-98e613451c6f","Type":"ContainerStarted","Data":"5bcd8603bc2d1790adb0fe6c925663c029661d4c38d11a36af38c808b3540a60"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.567744 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jl78z" podStartSLOduration=127.567720275 podStartE2EDuration="2m7.567720275s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:36.52423739 +0000 UTC m=+151.254748285" watchObservedRunningTime="2026-01-21 17:58:36.567720275 +0000 UTC m=+151.298231160" Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.583811 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cn7qm" event={"ID":"a58048c2-af3b-4ea3-820e-622c72edd91f","Type":"ContainerStarted","Data":"e043ffe3d050f497b3296277ad74f819d170f21950b595a6b7299624050bbc38"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.583874 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cn7qm" event={"ID":"a58048c2-af3b-4ea3-820e-622c72edd91f","Type":"ContainerStarted","Data":"659eb642918ef9ddff60fc087fe6fa7e04dec900bc744a726db393eec810f54c"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.590128 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:36 crc kubenswrapper[4808]: E0121 17:58:36.592340 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:37.09231845 +0000 UTC m=+151.822829345 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.601104 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-01-21 17:53:35 +0000 UTC, rotation deadline is 2026-10-30 00:05:04.712805222 +0000 UTC Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.601141 4808 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6750h6m28.111667517s for next certificate rotation Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.617807 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xcn7d" podStartSLOduration=127.617792952 podStartE2EDuration="2m7.617792952s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:36.583728246 +0000 UTC m=+151.314239141" watchObservedRunningTime="2026-01-21 17:58:36.617792952 +0000 UTC m=+151.348303837" Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.618038 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lq8cg" podStartSLOduration=127.618034312 podStartE2EDuration="2m7.618034312s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:36.615851212 +0000 UTC m=+151.346362107" watchObservedRunningTime="2026-01-21 17:58:36.618034312 +0000 UTC m=+151.348545197" Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.618806 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" event={"ID":"5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5","Type":"ContainerStarted","Data":"7d9f6eddf2262022cc1b232550ad1e5a53b71fd2cb73860ce8c91672eedcf75b"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.644510 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-s5khx" event={"ID":"e6e2b28f-7cfe-44a5-8a63-f17542f27cb6","Type":"ContainerStarted","Data":"67f877fede91f3aaec0fca716927cf06d43fbb71f60b680524d5aadbdf54d5f2"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.662623 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zjcb6" event={"ID":"044bc3f1-737c-4dda-9c24-d087ce039b42","Type":"ContainerStarted","Data":"4d3898a2281967a5a8afd0741b038409fc3dd7fda168722bd8bb8def8ac1bad2"} Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.666967 4808 patch_prober.go:28] interesting pod/downloads-7954f5f757-ftpvl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.667027 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ftpvl" podUID="196e1220-375f-4a05-90e0-ce571f606de4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.681625 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-2q8lz" Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.691319 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-k2nbx" Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.697314 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:36 crc kubenswrapper[4808]: E0121 17:58:36.697655 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:37.197641178 +0000 UTC m=+151.928152063 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.698534 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cn7qm" podStartSLOduration=127.698511344 podStartE2EDuration="2m7.698511344s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:36.695758831 +0000 UTC m=+151.426269726" watchObservedRunningTime="2026-01-21 17:58:36.698511344 +0000 UTC m=+151.429022229" Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.699000 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-blj6k" podStartSLOduration=127.698994514 podStartE2EDuration="2m7.698994514s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:36.660301397 +0000 UTC m=+151.390812292" watchObservedRunningTime="2026-01-21 17:58:36.698994514 +0000 UTC m=+151.429505399" Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.799616 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:36 crc kubenswrapper[4808]: E0121 17:58:36.802009 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:37.301991165 +0000 UTC m=+152.032502050 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:36 crc kubenswrapper[4808]: I0121 17:58:36.940691 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:36 crc kubenswrapper[4808]: E0121 17:58:36.961913 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:37.461893367 +0000 UTC m=+152.192404252 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.042812 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:37 crc kubenswrapper[4808]: E0121 17:58:37.043286 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:37.543265436 +0000 UTC m=+152.273776321 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.057500 4808 patch_prober.go:28] interesting pod/router-default-5444994796-xfblx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 17:58:37 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Jan 21 17:58:37 crc kubenswrapper[4808]: [+]process-running ok Jan 21 17:58:37 crc kubenswrapper[4808]: healthz check failed Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.057569 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xfblx" podUID="9a23bae9-c21c-40a6-aa8d-9c3524614dd4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.144575 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:37 crc kubenswrapper[4808]: E0121 17:58:37.145369 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:37.64535466 +0000 UTC m=+152.375865555 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.246058 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:37 crc kubenswrapper[4808]: E0121 17:58:37.246273 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:37.746231385 +0000 UTC m=+152.476742270 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.246460 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:37 crc kubenswrapper[4808]: E0121 17:58:37.246841 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:37.746829839 +0000 UTC m=+152.477340734 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.349682 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:37 crc kubenswrapper[4808]: E0121 17:58:37.349988 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:37.849972567 +0000 UTC m=+152.580483452 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.450889 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:37 crc kubenswrapper[4808]: E0121 17:58:37.451327 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:37.951311071 +0000 UTC m=+152.681821956 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.551918 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:37 crc kubenswrapper[4808]: E0121 17:58:37.552350 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:38.052329321 +0000 UTC m=+152.782840206 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.654277 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:37 crc kubenswrapper[4808]: E0121 17:58:37.654939 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:38.154916126 +0000 UTC m=+152.885427011 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.681306 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k" event={"ID":"d98f2d52-51f0-4843-8ba3-50264b3b3abe","Type":"ContainerStarted","Data":"d5aa7b4fcc7e9c1a1f2d99cb26def55a88d33444e3decafb60f6f0c8d4af85c8"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.681370 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k" event={"ID":"d98f2d52-51f0-4843-8ba3-50264b3b3abe","Type":"ContainerStarted","Data":"c32631826cde75984e7789fc9c91aca911f8de21a38632213a854fa6750f7d60"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.683857 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-s5khx" event={"ID":"e6e2b28f-7cfe-44a5-8a63-f17542f27cb6","Type":"ContainerStarted","Data":"6188dd6e6ed7b65082245d2c79c8a0e3acd78b518668486893a116c43f3e94bf"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.694883 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d2rkk" event={"ID":"c8a01e69-a842-4316-9fd6-a2850afa5b11","Type":"ContainerStarted","Data":"a99ccf6a0a4895eceabc47bde6cb5caac679cf484f60f6813aafee0266e5d269"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.694943 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d2rkk" event={"ID":"c8a01e69-a842-4316-9fd6-a2850afa5b11","Type":"ContainerStarted","Data":"500bac2a91b443b12da31bd729a9788cf46b88b3f84062946d4ab8c5bcabb7ef"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.717565 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zjcb6" event={"ID":"044bc3f1-737c-4dda-9c24-d087ce039b42","Type":"ContainerStarted","Data":"f23d657773216292106ed3aa99ff0bd1efe32905c5f40f7f088e6fe4611a22a8"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.717622 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zjcb6" event={"ID":"044bc3f1-737c-4dda-9c24-d087ce039b42","Type":"ContainerStarted","Data":"64718da62d6cd9dbad46fe9ef1455074a6c634bcfe5c7bb27345f03e97499367"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.718284 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-zjcb6" Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.735509 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z77hq" event={"ID":"d9f360c8-31a2-4577-8eec-d5d58c106933","Type":"ContainerStarted","Data":"d057c3b3b389a03d88cdc246d71d044e89e442ec55ba9cebd9b7be70263a6a46"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.749593 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-79g2n" event={"ID":"29185a07-7266-4975-870b-127f880f2cf6","Type":"ContainerStarted","Data":"ee2d138bd73d5ecf1b0dd5497452acec943c9cc34d3b77a362c0f73bb02e559d"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.758377 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.761902 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mqm6k" podStartSLOduration=128.761883782 podStartE2EDuration="2m8.761883782s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:37.720213192 +0000 UTC m=+152.450724077" watchObservedRunningTime="2026-01-21 17:58:37.761883782 +0000 UTC m=+152.492394677" Jan 21 17:58:37 crc kubenswrapper[4808]: E0121 17:58:37.763613 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:38.263596073 +0000 UTC m=+152.994106958 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.767333 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2" event={"ID":"c0b6ae2d-bfaf-4535-9906-31b19d4fe489","Type":"ContainerStarted","Data":"38c8bccfb1c4e3a455b6493d4db9ce026347b6cf3b09d5447ebf3fdd0d50db88"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.775581 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cn7qm" event={"ID":"a58048c2-af3b-4ea3-820e-622c72edd91f","Type":"ContainerStarted","Data":"0dd41f8b2a3e66f7df34dc0ed39e365b842147ed2de0d1d53ecff166b48ea31f"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.811900 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q8wfq" event={"ID":"bc0b3e27-43a2-4a5b-98bc-d371958b2bbf","Type":"ContainerStarted","Data":"83fec74deb63c806793552dd2e60c60188a3f5a68c92d9ea43a4bf40ddae31bb"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.811958 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q8wfq" event={"ID":"bc0b3e27-43a2-4a5b-98bc-d371958b2bbf","Type":"ContainerStarted","Data":"432ac8e89da3ca96dd1ceb51ffde0818d12f4de57a47eabf5ad8ee3d1d665c79"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.813590 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-zjcb6" podStartSLOduration=8.813579256 podStartE2EDuration="8.813579256s" podCreationTimestamp="2026-01-21 17:58:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:37.763389084 +0000 UTC m=+152.493899979" watchObservedRunningTime="2026-01-21 17:58:37.813579256 +0000 UTC m=+152.544090141" Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.813700 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q8wfq" Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.814165 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-s5khx" podStartSLOduration=128.81415984 podStartE2EDuration="2m8.81415984s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:37.803059582 +0000 UTC m=+152.533570467" watchObservedRunningTime="2026-01-21 17:58:37.81415984 +0000 UTC m=+152.544670725" Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.821684 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wj246" event={"ID":"030a3f02-a9cd-4a3c-890d-686e2b7df274","Type":"ContainerStarted","Data":"2020979e55df706dbd6f7fd9a5cd8cc67ee8f19e75daf54b6b3151649b517a0d"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.825918 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh" event={"ID":"77ae37d7-ba3c-4b77-85cb-f7cb648ab795","Type":"ContainerStarted","Data":"46cefd42b756af9f8bd2933fde0e14de33c4b7f4fd7485d89922c1966c1b714c"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.826858 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh" Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.838824 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4gg7j" event={"ID":"c6350524-fc0a-4bbc-b803-cea5fd8dc76c","Type":"ContainerStarted","Data":"22817a8e7e23484a8a4104fa9be4cff6d33701c288626fa2b422f0f94cca725f"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.839800 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4gg7j" Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.841489 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d2rkk" podStartSLOduration=128.841466778 podStartE2EDuration="2m8.841466778s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:37.840019998 +0000 UTC m=+152.570530883" watchObservedRunningTime="2026-01-21 17:58:37.841466778 +0000 UTC m=+152.571977663" Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.859883 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:37 crc kubenswrapper[4808]: E0121 17:58:37.862066 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:38.362049577 +0000 UTC m=+153.092560472 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.863834 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv" event={"ID":"e897d5b0-11f6-4e3f-99d2-22a7d093738a","Type":"ContainerStarted","Data":"b105bc7644bff6c076dd40a5b468a2b08858bdd4c41d04814bf26483df88f2a6"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.887900 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-79g2n" podStartSLOduration=7.887877974 podStartE2EDuration="7.887877974s" podCreationTimestamp="2026-01-21 17:58:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:37.88415139 +0000 UTC m=+152.614662285" watchObservedRunningTime="2026-01-21 17:58:37.887877974 +0000 UTC m=+152.618388859" Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.890031 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4gg7j" Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.894632 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-jxgs8" event={"ID":"b5ad8de7-53f2-475d-be7b-acb85ad7f53a","Type":"ContainerStarted","Data":"ace18d4901bf7c712b99308ff8ae1b1b3c306d6b1cb4796540277d99e54b054a"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.901721 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jmp6c" event={"ID":"2c1a10ce-320b-48fe-9eca-ab5e6d110441","Type":"ContainerStarted","Data":"563d542a7bc40c04e26d88efb8f8339cc84a07bdc9e68aed424ae351a67becd7"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.902659 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jmp6c" Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.915883 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" event={"ID":"5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5","Type":"ContainerStarted","Data":"cbd58dac49caa9a8cf3d13700bf902392cd9d8cbe2e1cb48e8dbe6e67b4cf935"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.916799 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.918917 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-slth9" event={"ID":"433911de-428a-4261-8680-66a30c308919","Type":"ContainerStarted","Data":"3a24788e8014b51677ecb2aadc700590218b75b90cf3723ed3edf1a6ba8fe34f"} Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.940183 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh" podStartSLOduration=128.940158902 podStartE2EDuration="2m8.940158902s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:37.928509311 +0000 UTC m=+152.659020196" watchObservedRunningTime="2026-01-21 17:58:37.940158902 +0000 UTC m=+152.670669787" Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.942641 4808 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-jmp6c container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.942698 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jmp6c" podUID="2c1a10ce-320b-48fe-9eca-ab5e6d110441" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.953430 4808 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-pvk7n container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.953500 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" podUID="5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Jan 21 17:58:37 crc kubenswrapper[4808]: I0121 17:58:37.964769 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:37 crc kubenswrapper[4808]: E0121 17:58:37.966761 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:38.466735199 +0000 UTC m=+153.197246084 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.031990 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-z77hq" podStartSLOduration=129.031968042 podStartE2EDuration="2m9.031968042s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:37.993591138 +0000 UTC m=+152.724102023" watchObservedRunningTime="2026-01-21 17:58:38.031968042 +0000 UTC m=+152.762478927" Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.039452 4808 patch_prober.go:28] interesting pod/router-default-5444994796-xfblx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 17:58:38 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Jan 21 17:58:38 crc kubenswrapper[4808]: [+]process-running ok Jan 21 17:58:38 crc kubenswrapper[4808]: healthz check failed Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.039519 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xfblx" podUID="9a23bae9-c21c-40a6-aa8d-9c3524614dd4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.068010 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:38 crc kubenswrapper[4808]: E0121 17:58:38.069477 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:38.5694594 +0000 UTC m=+153.299970285 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.102614 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wj246" podStartSLOduration=129.102579857 podStartE2EDuration="2m9.102579857s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:38.093054664 +0000 UTC m=+152.823565559" watchObservedRunningTime="2026-01-21 17:58:38.102579857 +0000 UTC m=+152.833090762" Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.104951 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2" podStartSLOduration=129.104933774 podStartE2EDuration="2m9.104933774s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:38.055268034 +0000 UTC m=+152.785778929" watchObservedRunningTime="2026-01-21 17:58:38.104933774 +0000 UTC m=+152.835444649" Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.134838 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-p9d2b" Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.139541 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4gg7j" podStartSLOduration=129.139520852 podStartE2EDuration="2m9.139520852s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:38.130076502 +0000 UTC m=+152.860587397" watchObservedRunningTime="2026-01-21 17:58:38.139520852 +0000 UTC m=+152.870031737" Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.165715 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q8wfq" podStartSLOduration=129.165698893 podStartE2EDuration="2m9.165698893s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:38.165492224 +0000 UTC m=+152.896003119" watchObservedRunningTime="2026-01-21 17:58:38.165698893 +0000 UTC m=+152.896209778" Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.171568 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:38 crc kubenswrapper[4808]: E0121 17:58:38.172340 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:38.672322546 +0000 UTC m=+153.402833431 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.227896 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" podStartSLOduration=129.22788174 podStartE2EDuration="2m9.22788174s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:38.226049754 +0000 UTC m=+152.956560639" watchObservedRunningTime="2026-01-21 17:58:38.22788174 +0000 UTC m=+152.958392625" Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.265881 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jmp6c" podStartSLOduration=129.265864068 podStartE2EDuration="2m9.265864068s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:38.263613095 +0000 UTC m=+152.994123990" watchObservedRunningTime="2026-01-21 17:58:38.265864068 +0000 UTC m=+152.996374953" Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.274818 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:38 crc kubenswrapper[4808]: E0121 17:58:38.275200 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:38.775187443 +0000 UTC m=+153.505698328 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.312794 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6tbv" podStartSLOduration=129.312771064 podStartE2EDuration="2m9.312771064s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:38.312677941 +0000 UTC m=+153.043188836" watchObservedRunningTime="2026-01-21 17:58:38.312771064 +0000 UTC m=+153.043281949" Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.375688 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:38 crc kubenswrapper[4808]: E0121 17:58:38.376075 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:38.876060187 +0000 UTC m=+153.606571072 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.434792 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qfggh" Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.476776 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:38 crc kubenswrapper[4808]: E0121 17:58:38.477121 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:38.977105028 +0000 UTC m=+153.707615913 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.578014 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:38 crc kubenswrapper[4808]: E0121 17:58:38.578252 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:39.078207712 +0000 UTC m=+153.808718607 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.578535 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:38 crc kubenswrapper[4808]: E0121 17:58:38.578869 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:39.078853849 +0000 UTC m=+153.809364734 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.679301 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:38 crc kubenswrapper[4808]: E0121 17:58:38.679598 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:39.179582237 +0000 UTC m=+153.910093122 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.780205 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:38 crc kubenswrapper[4808]: E0121 17:58:38.780823 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:39.280811106 +0000 UTC m=+154.011321991 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.881735 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:38 crc kubenswrapper[4808]: E0121 17:58:38.882024 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:39.381993613 +0000 UTC m=+154.112504508 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.882173 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:38 crc kubenswrapper[4808]: E0121 17:58:38.882556 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:39.382541026 +0000 UTC m=+154.113051921 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.943555 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-glhp5" event={"ID":"922aad90-2298-4afb-816e-063c73f18c6d","Type":"ContainerStarted","Data":"86bd0b75b8d461d0deaf4bd2d88d827cc46717c21c5bbebad812086c82d83efe"} Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.943609 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-glhp5" event={"ID":"922aad90-2298-4afb-816e-063c73f18c6d","Type":"ContainerStarted","Data":"5238408c415914f727ffd73019eedb9d5e33b2d24d3364a83e4aaa9c0650a33f"} Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.945997 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z77hq" event={"ID":"d9f360c8-31a2-4577-8eec-d5d58c106933","Type":"ContainerStarted","Data":"84c4fce47063bc50fd3b6ad6f0d2815336638137403b55cb87523709a791bb7c"} Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.948175 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-jxgs8" event={"ID":"b5ad8de7-53f2-475d-be7b-acb85ad7f53a","Type":"ContainerStarted","Data":"3421120fffc8e9b79856e73282df00b380818120952455bdff3df59112170f62"} Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.949676 4808 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-pvk7n container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.949725 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" podUID="5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.959212 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jmp6c" Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.988898 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:38 crc kubenswrapper[4808]: E0121 17:58:38.988972 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:39.488957529 +0000 UTC m=+154.219468414 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:38 crc kubenswrapper[4808]: I0121 17:58:38.990377 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:38 crc kubenswrapper[4808]: E0121 17:58:38.991143 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:39.491120528 +0000 UTC m=+154.221631483 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.037696 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-jxgs8" podStartSLOduration=130.0376723 podStartE2EDuration="2m10.0376723s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:38.988199968 +0000 UTC m=+153.718710853" watchObservedRunningTime="2026-01-21 17:58:39.0376723 +0000 UTC m=+153.768183185" Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.040498 4808 patch_prober.go:28] interesting pod/router-default-5444994796-xfblx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 17:58:39 crc kubenswrapper[4808]: [-]has-synced failed: reason withheld Jan 21 17:58:39 crc kubenswrapper[4808]: [+]process-running ok Jan 21 17:58:39 crc kubenswrapper[4808]: healthz check failed Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.040578 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xfblx" podUID="9a23bae9-c21c-40a6-aa8d-9c3524614dd4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.094702 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:39 crc kubenswrapper[4808]: E0121 17:58:39.094859 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:39.59484171 +0000 UTC m=+154.325352595 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.095180 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:39 crc kubenswrapper[4808]: E0121 17:58:39.099621 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:39.599611087 +0000 UTC m=+154.330121962 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.123905 4808 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.196075 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:39 crc kubenswrapper[4808]: E0121 17:58:39.196292 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:39.696264597 +0000 UTC m=+154.426775482 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.297663 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:39 crc kubenswrapper[4808]: E0121 17:58:39.298045 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:39.798027898 +0000 UTC m=+154.528538783 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.398606 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:39 crc kubenswrapper[4808]: E0121 17:58:39.398821 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:39.898785158 +0000 UTC m=+154.629296053 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.499696 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:39 crc kubenswrapper[4808]: E0121 17:58:39.500184 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:40.000164193 +0000 UTC m=+154.730675128 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.600155 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:39 crc kubenswrapper[4808]: E0121 17:58:39.600374 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:40.100345969 +0000 UTC m=+154.830856854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.701474 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:39 crc kubenswrapper[4808]: E0121 17:58:39.701848 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:40.201832959 +0000 UTC m=+154.932343844 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.720068 4808 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-01-21T17:58:39.123932451Z","Handler":null,"Name":""} Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.802206 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:39 crc kubenswrapper[4808]: E0121 17:58:39.802405 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:40.302374919 +0000 UTC m=+155.032885804 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.802549 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:39 crc kubenswrapper[4808]: E0121 17:58:39.802888 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:40.30287672 +0000 UTC m=+155.033387605 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.878221 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hfz2d"] Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.880138 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hfz2d" Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.882622 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.893182 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hfz2d"] Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.903141 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:39 crc kubenswrapper[4808]: E0121 17:58:39.903457 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 17:58:40.403423651 +0000 UTC m=+155.133934536 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.903988 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:39 crc kubenswrapper[4808]: E0121 17:58:39.904357 4808 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 17:58:40.404348999 +0000 UTC m=+155.134859884 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-n56cl" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.954691 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-glhp5" event={"ID":"922aad90-2298-4afb-816e-063c73f18c6d","Type":"ContainerStarted","Data":"a39bcfffb83aaea6a25ef3aa8a21ac5e3830ecb1516dea6261367f6fdfc5097d"} Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.956171 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.968098 4808 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Jan 21 17:58:39 crc kubenswrapper[4808]: I0121 17:58:39.968141 4808 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.006119 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.006437 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09939aec-8ef7-4b4d-a934-9f00e19bc7ed-catalog-content\") pod \"community-operators-hfz2d\" (UID: \"09939aec-8ef7-4b4d-a934-9f00e19bc7ed\") " pod="openshift-marketplace/community-operators-hfz2d" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.006467 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09939aec-8ef7-4b4d-a934-9f00e19bc7ed-utilities\") pod \"community-operators-hfz2d\" (UID: \"09939aec-8ef7-4b4d-a934-9f00e19bc7ed\") " pod="openshift-marketplace/community-operators-hfz2d" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.006495 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mztfc\" (UniqueName: \"kubernetes.io/projected/09939aec-8ef7-4b4d-a934-9f00e19bc7ed-kube-api-access-mztfc\") pod \"community-operators-hfz2d\" (UID: \"09939aec-8ef7-4b4d-a934-9f00e19bc7ed\") " pod="openshift-marketplace/community-operators-hfz2d" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.023777 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.041690 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.043263 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.046602 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-xfblx" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.084118 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2lzcv"] Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.085100 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2lzcv" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.087506 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.103038 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2lzcv"] Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.107373 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.107476 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09939aec-8ef7-4b4d-a934-9f00e19bc7ed-catalog-content\") pod \"community-operators-hfz2d\" (UID: \"09939aec-8ef7-4b4d-a934-9f00e19bc7ed\") " pod="openshift-marketplace/community-operators-hfz2d" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.107497 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09939aec-8ef7-4b4d-a934-9f00e19bc7ed-utilities\") pod \"community-operators-hfz2d\" (UID: \"09939aec-8ef7-4b4d-a934-9f00e19bc7ed\") " pod="openshift-marketplace/community-operators-hfz2d" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.107533 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mztfc\" (UniqueName: \"kubernetes.io/projected/09939aec-8ef7-4b4d-a934-9f00e19bc7ed-kube-api-access-mztfc\") pod \"community-operators-hfz2d\" (UID: \"09939aec-8ef7-4b4d-a934-9f00e19bc7ed\") " pod="openshift-marketplace/community-operators-hfz2d" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.109017 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09939aec-8ef7-4b4d-a934-9f00e19bc7ed-catalog-content\") pod \"community-operators-hfz2d\" (UID: \"09939aec-8ef7-4b4d-a934-9f00e19bc7ed\") " pod="openshift-marketplace/community-operators-hfz2d" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.109260 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09939aec-8ef7-4b4d-a934-9f00e19bc7ed-utilities\") pod \"community-operators-hfz2d\" (UID: \"09939aec-8ef7-4b4d-a934-9f00e19bc7ed\") " pod="openshift-marketplace/community-operators-hfz2d" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.118015 4808 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.118061 4808 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.142308 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mztfc\" (UniqueName: \"kubernetes.io/projected/09939aec-8ef7-4b4d-a934-9f00e19bc7ed-kube-api-access-mztfc\") pod \"community-operators-hfz2d\" (UID: \"09939aec-8ef7-4b4d-a934-9f00e19bc7ed\") " pod="openshift-marketplace/community-operators-hfz2d" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.168916 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.198196 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hfz2d" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.208119 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6f5w\" (UniqueName: \"kubernetes.io/projected/6bb3e28c-45c6-4992-beee-69dfd1ae7d55-kube-api-access-m6f5w\") pod \"certified-operators-2lzcv\" (UID: \"6bb3e28c-45c6-4992-beee-69dfd1ae7d55\") " pod="openshift-marketplace/certified-operators-2lzcv" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.208190 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bb3e28c-45c6-4992-beee-69dfd1ae7d55-utilities\") pod \"certified-operators-2lzcv\" (UID: \"6bb3e28c-45c6-4992-beee-69dfd1ae7d55\") " pod="openshift-marketplace/certified-operators-2lzcv" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.208227 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bb3e28c-45c6-4992-beee-69dfd1ae7d55-catalog-content\") pod \"certified-operators-2lzcv\" (UID: \"6bb3e28c-45c6-4992-beee-69dfd1ae7d55\") " pod="openshift-marketplace/certified-operators-2lzcv" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.209971 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-n56cl\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.308988 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bb3e28c-45c6-4992-beee-69dfd1ae7d55-utilities\") pod \"certified-operators-2lzcv\" (UID: \"6bb3e28c-45c6-4992-beee-69dfd1ae7d55\") " pod="openshift-marketplace/certified-operators-2lzcv" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.309050 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bb3e28c-45c6-4992-beee-69dfd1ae7d55-catalog-content\") pod \"certified-operators-2lzcv\" (UID: \"6bb3e28c-45c6-4992-beee-69dfd1ae7d55\") " pod="openshift-marketplace/certified-operators-2lzcv" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.309131 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6f5w\" (UniqueName: \"kubernetes.io/projected/6bb3e28c-45c6-4992-beee-69dfd1ae7d55-kube-api-access-m6f5w\") pod \"certified-operators-2lzcv\" (UID: \"6bb3e28c-45c6-4992-beee-69dfd1ae7d55\") " pod="openshift-marketplace/certified-operators-2lzcv" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.309618 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bb3e28c-45c6-4992-beee-69dfd1ae7d55-utilities\") pod \"certified-operators-2lzcv\" (UID: \"6bb3e28c-45c6-4992-beee-69dfd1ae7d55\") " pod="openshift-marketplace/certified-operators-2lzcv" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.309677 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bb3e28c-45c6-4992-beee-69dfd1ae7d55-catalog-content\") pod \"certified-operators-2lzcv\" (UID: \"6bb3e28c-45c6-4992-beee-69dfd1ae7d55\") " pod="openshift-marketplace/certified-operators-2lzcv" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.325427 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nr5dg"] Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.326564 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nr5dg" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.360075 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6f5w\" (UniqueName: \"kubernetes.io/projected/6bb3e28c-45c6-4992-beee-69dfd1ae7d55-kube-api-access-m6f5w\") pod \"certified-operators-2lzcv\" (UID: \"6bb3e28c-45c6-4992-beee-69dfd1ae7d55\") " pod="openshift-marketplace/certified-operators-2lzcv" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.396560 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nr5dg"] Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.417449 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2lzcv" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.418788 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6221711c-1051-44e9-8fad-029c73d0ae71-catalog-content\") pod \"community-operators-nr5dg\" (UID: \"6221711c-1051-44e9-8fad-029c73d0ae71\") " pod="openshift-marketplace/community-operators-nr5dg" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.418882 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xzbc\" (UniqueName: \"kubernetes.io/projected/6221711c-1051-44e9-8fad-029c73d0ae71-kube-api-access-7xzbc\") pod \"community-operators-nr5dg\" (UID: \"6221711c-1051-44e9-8fad-029c73d0ae71\") " pod="openshift-marketplace/community-operators-nr5dg" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.418908 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6221711c-1051-44e9-8fad-029c73d0ae71-utilities\") pod \"community-operators-nr5dg\" (UID: \"6221711c-1051-44e9-8fad-029c73d0ae71\") " pod="openshift-marketplace/community-operators-nr5dg" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.485615 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.509051 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4fsv9"] Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.509957 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4fsv9" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.523281 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a84f07d-c20d-4a1d-a272-118d29bcdb60-catalog-content\") pod \"certified-operators-4fsv9\" (UID: \"3a84f07d-c20d-4a1d-a272-118d29bcdb60\") " pod="openshift-marketplace/certified-operators-4fsv9" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.523372 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a84f07d-c20d-4a1d-a272-118d29bcdb60-utilities\") pod \"certified-operators-4fsv9\" (UID: \"3a84f07d-c20d-4a1d-a272-118d29bcdb60\") " pod="openshift-marketplace/certified-operators-4fsv9" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.523406 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6221711c-1051-44e9-8fad-029c73d0ae71-catalog-content\") pod \"community-operators-nr5dg\" (UID: \"6221711c-1051-44e9-8fad-029c73d0ae71\") " pod="openshift-marketplace/community-operators-nr5dg" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.523455 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xzbc\" (UniqueName: \"kubernetes.io/projected/6221711c-1051-44e9-8fad-029c73d0ae71-kube-api-access-7xzbc\") pod \"community-operators-nr5dg\" (UID: \"6221711c-1051-44e9-8fad-029c73d0ae71\") " pod="openshift-marketplace/community-operators-nr5dg" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.523475 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6221711c-1051-44e9-8fad-029c73d0ae71-utilities\") pod \"community-operators-nr5dg\" (UID: \"6221711c-1051-44e9-8fad-029c73d0ae71\") " pod="openshift-marketplace/community-operators-nr5dg" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.523493 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tggj\" (UniqueName: \"kubernetes.io/projected/3a84f07d-c20d-4a1d-a272-118d29bcdb60-kube-api-access-6tggj\") pod \"certified-operators-4fsv9\" (UID: \"3a84f07d-c20d-4a1d-a272-118d29bcdb60\") " pod="openshift-marketplace/certified-operators-4fsv9" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.532515 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4fsv9"] Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.537086 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6221711c-1051-44e9-8fad-029c73d0ae71-catalog-content\") pod \"community-operators-nr5dg\" (UID: \"6221711c-1051-44e9-8fad-029c73d0ae71\") " pod="openshift-marketplace/community-operators-nr5dg" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.537468 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6221711c-1051-44e9-8fad-029c73d0ae71-utilities\") pod \"community-operators-nr5dg\" (UID: \"6221711c-1051-44e9-8fad-029c73d0ae71\") " pod="openshift-marketplace/community-operators-nr5dg" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.568505 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xzbc\" (UniqueName: \"kubernetes.io/projected/6221711c-1051-44e9-8fad-029c73d0ae71-kube-api-access-7xzbc\") pod \"community-operators-nr5dg\" (UID: \"6221711c-1051-44e9-8fad-029c73d0ae71\") " pod="openshift-marketplace/community-operators-nr5dg" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.615307 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hfz2d"] Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.624750 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tggj\" (UniqueName: \"kubernetes.io/projected/3a84f07d-c20d-4a1d-a272-118d29bcdb60-kube-api-access-6tggj\") pod \"certified-operators-4fsv9\" (UID: \"3a84f07d-c20d-4a1d-a272-118d29bcdb60\") " pod="openshift-marketplace/certified-operators-4fsv9" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.624805 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a84f07d-c20d-4a1d-a272-118d29bcdb60-catalog-content\") pod \"certified-operators-4fsv9\" (UID: \"3a84f07d-c20d-4a1d-a272-118d29bcdb60\") " pod="openshift-marketplace/certified-operators-4fsv9" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.624834 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a84f07d-c20d-4a1d-a272-118d29bcdb60-utilities\") pod \"certified-operators-4fsv9\" (UID: \"3a84f07d-c20d-4a1d-a272-118d29bcdb60\") " pod="openshift-marketplace/certified-operators-4fsv9" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.625269 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a84f07d-c20d-4a1d-a272-118d29bcdb60-utilities\") pod \"certified-operators-4fsv9\" (UID: \"3a84f07d-c20d-4a1d-a272-118d29bcdb60\") " pod="openshift-marketplace/certified-operators-4fsv9" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.626465 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a84f07d-c20d-4a1d-a272-118d29bcdb60-catalog-content\") pod \"certified-operators-4fsv9\" (UID: \"3a84f07d-c20d-4a1d-a272-118d29bcdb60\") " pod="openshift-marketplace/certified-operators-4fsv9" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.665034 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tggj\" (UniqueName: \"kubernetes.io/projected/3a84f07d-c20d-4a1d-a272-118d29bcdb60-kube-api-access-6tggj\") pod \"certified-operators-4fsv9\" (UID: \"3a84f07d-c20d-4a1d-a272-118d29bcdb60\") " pod="openshift-marketplace/certified-operators-4fsv9" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.673021 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nr5dg" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.845053 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4fsv9" Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.950368 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-n56cl"] Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.966073 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-glhp5" event={"ID":"922aad90-2298-4afb-816e-063c73f18c6d","Type":"ContainerStarted","Data":"fff0f8103a3748a44bbd4f427436c31a46663292d9036a0fc301fcff14810c55"} Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.968991 4808 generic.go:334] "Generic (PLEG): container finished" podID="c0b6ae2d-bfaf-4535-9906-31b19d4fe489" containerID="38c8bccfb1c4e3a455b6493d4db9ce026347b6cf3b09d5447ebf3fdd0d50db88" exitCode=0 Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.969053 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2" event={"ID":"c0b6ae2d-bfaf-4535-9906-31b19d4fe489","Type":"ContainerDied","Data":"38c8bccfb1c4e3a455b6493d4db9ce026347b6cf3b09d5447ebf3fdd0d50db88"} Jan 21 17:58:40 crc kubenswrapper[4808]: I0121 17:58:40.988799 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-glhp5" podStartSLOduration=11.988776646 podStartE2EDuration="11.988776646s" podCreationTimestamp="2026-01-21 17:58:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:40.988644671 +0000 UTC m=+155.719155556" watchObservedRunningTime="2026-01-21 17:58:40.988776646 +0000 UTC m=+155.719287531" Jan 21 17:58:41 crc kubenswrapper[4808]: I0121 17:58:40.998082 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfz2d" event={"ID":"09939aec-8ef7-4b4d-a934-9f00e19bc7ed","Type":"ContainerStarted","Data":"d25da3ea931b69f3f101613565c2596843abdaa238cd7b071f8a804bb697956c"} Jan 21 17:58:41 crc kubenswrapper[4808]: I0121 17:58:40.998132 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfz2d" event={"ID":"09939aec-8ef7-4b4d-a934-9f00e19bc7ed","Type":"ContainerStarted","Data":"5cdc40ff058c140c7aa79d59e5c3b7bdaea72112b20f4d6beb57b31bb9f1dd07"} Jan 21 17:58:41 crc kubenswrapper[4808]: I0121 17:58:41.000329 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 17:58:41 crc kubenswrapper[4808]: I0121 17:58:41.027333 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nr5dg"] Jan 21 17:58:41 crc kubenswrapper[4808]: I0121 17:58:41.085002 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2lzcv"] Jan 21 17:58:41 crc kubenswrapper[4808]: I0121 17:58:41.129722 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4fsv9"] Jan 21 17:58:41 crc kubenswrapper[4808]: W0121 17:58:41.142792 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a84f07d_c20d_4a1d_a272_118d29bcdb60.slice/crio-4661d49dbe78a3ff03c5f1fcc3258eb33d61c9501d34d336285b2760c00bfc35 WatchSource:0}: Error finding container 4661d49dbe78a3ff03c5f1fcc3258eb33d61c9501d34d336285b2760c00bfc35: Status 404 returned error can't find the container with id 4661d49dbe78a3ff03c5f1fcc3258eb33d61c9501d34d336285b2760c00bfc35 Jan 21 17:58:41 crc kubenswrapper[4808]: I0121 17:58:41.528629 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Jan 21 17:58:41 crc kubenswrapper[4808]: I0121 17:58:41.876016 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5z654"] Jan 21 17:58:41 crc kubenswrapper[4808]: I0121 17:58:41.877107 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5z654" Jan 21 17:58:41 crc kubenswrapper[4808]: I0121 17:58:41.881043 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 21 17:58:41 crc kubenswrapper[4808]: I0121 17:58:41.887045 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5z654"] Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.000282 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.001277 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.006236 4808 patch_prober.go:28] interesting pod/console-f9d7485db-h6gm7 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.006311 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-h6gm7" podUID="05688b08-2c9f-4c6c-967a-ff64de7837c0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.014460 4808 generic.go:334] "Generic (PLEG): container finished" podID="6221711c-1051-44e9-8fad-029c73d0ae71" containerID="97ffdcaab65867b1dd6ac8390bb30fc9d2cc128c836e8f1e0bfd561a53ff6f0a" exitCode=0 Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.014530 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nr5dg" event={"ID":"6221711c-1051-44e9-8fad-029c73d0ae71","Type":"ContainerDied","Data":"97ffdcaab65867b1dd6ac8390bb30fc9d2cc128c836e8f1e0bfd561a53ff6f0a"} Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.014562 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nr5dg" event={"ID":"6221711c-1051-44e9-8fad-029c73d0ae71","Type":"ContainerStarted","Data":"f1b07c4835ad0926217706d2ddf4498515cc5ff0bf5cd300ef2894226ce471f7"} Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.025094 4808 generic.go:334] "Generic (PLEG): container finished" podID="09939aec-8ef7-4b4d-a934-9f00e19bc7ed" containerID="d25da3ea931b69f3f101613565c2596843abdaa238cd7b071f8a804bb697956c" exitCode=0 Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.025209 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfz2d" event={"ID":"09939aec-8ef7-4b4d-a934-9f00e19bc7ed","Type":"ContainerDied","Data":"d25da3ea931b69f3f101613565c2596843abdaa238cd7b071f8a804bb697956c"} Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.030044 4808 generic.go:334] "Generic (PLEG): container finished" podID="6bb3e28c-45c6-4992-beee-69dfd1ae7d55" containerID="5d7a6b63cf66a4c74602ce95aac4a7cdbebba8d0d5c653c14a45cda28984c7e4" exitCode=0 Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.030994 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lzcv" event={"ID":"6bb3e28c-45c6-4992-beee-69dfd1ae7d55","Type":"ContainerDied","Data":"5d7a6b63cf66a4c74602ce95aac4a7cdbebba8d0d5c653c14a45cda28984c7e4"} Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.031027 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lzcv" event={"ID":"6bb3e28c-45c6-4992-beee-69dfd1ae7d55","Type":"ContainerStarted","Data":"ed0f63879b8d9fda79a6c605be730962090c2c91ed5a5d5813f707491810975a"} Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.034479 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" event={"ID":"2b3e892c-c86a-42cd-b257-d722cd033b5e","Type":"ContainerStarted","Data":"db91f40f6def2d0a37ace9509c0e5d78a6880ff28cc615fa2f892df12ac32746"} Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.034523 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" event={"ID":"2b3e892c-c86a-42cd-b257-d722cd033b5e","Type":"ContainerStarted","Data":"d2434840d85f0c1ca009b377e0f76661a76a8aef8c14082e80d506bd909910e6"} Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.035150 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.037011 4808 generic.go:334] "Generic (PLEG): container finished" podID="3a84f07d-c20d-4a1d-a272-118d29bcdb60" containerID="8032d1838a45c097010150327581f76f18bcef8e58d62811ebccf52728ff3f8f" exitCode=0 Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.037896 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4fsv9" event={"ID":"3a84f07d-c20d-4a1d-a272-118d29bcdb60","Type":"ContainerDied","Data":"8032d1838a45c097010150327581f76f18bcef8e58d62811ebccf52728ff3f8f"} Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.037933 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4fsv9" event={"ID":"3a84f07d-c20d-4a1d-a272-118d29bcdb60","Type":"ContainerStarted","Data":"4661d49dbe78a3ff03c5f1fcc3258eb33d61c9501d34d336285b2760c00bfc35"} Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.047780 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/595fab48-bf4b-443f-9367-d5718a4672e3-catalog-content\") pod \"redhat-marketplace-5z654\" (UID: \"595fab48-bf4b-443f-9367-d5718a4672e3\") " pod="openshift-marketplace/redhat-marketplace-5z654" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.047860 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clswl\" (UniqueName: \"kubernetes.io/projected/595fab48-bf4b-443f-9367-d5718a4672e3-kube-api-access-clswl\") pod \"redhat-marketplace-5z654\" (UID: \"595fab48-bf4b-443f-9367-d5718a4672e3\") " pod="openshift-marketplace/redhat-marketplace-5z654" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.047932 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/595fab48-bf4b-443f-9367-d5718a4672e3-utilities\") pod \"redhat-marketplace-5z654\" (UID: \"595fab48-bf4b-443f-9367-d5718a4672e3\") " pod="openshift-marketplace/redhat-marketplace-5z654" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.101740 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" podStartSLOduration=133.101711381 podStartE2EDuration="2m13.101711381s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:42.093751923 +0000 UTC m=+156.824262828" watchObservedRunningTime="2026-01-21 17:58:42.101711381 +0000 UTC m=+156.832222276" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.149910 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clswl\" (UniqueName: \"kubernetes.io/projected/595fab48-bf4b-443f-9367-d5718a4672e3-kube-api-access-clswl\") pod \"redhat-marketplace-5z654\" (UID: \"595fab48-bf4b-443f-9367-d5718a4672e3\") " pod="openshift-marketplace/redhat-marketplace-5z654" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.150017 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/595fab48-bf4b-443f-9367-d5718a4672e3-utilities\") pod \"redhat-marketplace-5z654\" (UID: \"595fab48-bf4b-443f-9367-d5718a4672e3\") " pod="openshift-marketplace/redhat-marketplace-5z654" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.150170 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/595fab48-bf4b-443f-9367-d5718a4672e3-catalog-content\") pod \"redhat-marketplace-5z654\" (UID: \"595fab48-bf4b-443f-9367-d5718a4672e3\") " pod="openshift-marketplace/redhat-marketplace-5z654" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.153822 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/595fab48-bf4b-443f-9367-d5718a4672e3-catalog-content\") pod \"redhat-marketplace-5z654\" (UID: \"595fab48-bf4b-443f-9367-d5718a4672e3\") " pod="openshift-marketplace/redhat-marketplace-5z654" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.155210 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/595fab48-bf4b-443f-9367-d5718a4672e3-utilities\") pod \"redhat-marketplace-5z654\" (UID: \"595fab48-bf4b-443f-9367-d5718a4672e3\") " pod="openshift-marketplace/redhat-marketplace-5z654" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.180494 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clswl\" (UniqueName: \"kubernetes.io/projected/595fab48-bf4b-443f-9367-d5718a4672e3-kube-api-access-clswl\") pod \"redhat-marketplace-5z654\" (UID: \"595fab48-bf4b-443f-9367-d5718a4672e3\") " pod="openshift-marketplace/redhat-marketplace-5z654" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.192779 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5z654" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.250447 4808 patch_prober.go:28] interesting pod/downloads-7954f5f757-ftpvl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.250504 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ftpvl" podUID="196e1220-375f-4a05-90e0-ce571f606de4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.251165 4808 patch_prober.go:28] interesting pod/downloads-7954f5f757-ftpvl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.251214 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-ftpvl" podUID="196e1220-375f-4a05-90e0-ce571f606de4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.298374 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-79r4z"] Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.302207 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-79r4z" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.316715 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-79r4z"] Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.419788 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.456938 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0691f390-5c98-4c03-9839-e72ba2e6c947-catalog-content\") pod \"redhat-marketplace-79r4z\" (UID: \"0691f390-5c98-4c03-9839-e72ba2e6c947\") " pod="openshift-marketplace/redhat-marketplace-79r4z" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.456999 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjb9v\" (UniqueName: \"kubernetes.io/projected/0691f390-5c98-4c03-9839-e72ba2e6c947-kube-api-access-rjb9v\") pod \"redhat-marketplace-79r4z\" (UID: \"0691f390-5c98-4c03-9839-e72ba2e6c947\") " pod="openshift-marketplace/redhat-marketplace-79r4z" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.457091 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0691f390-5c98-4c03-9839-e72ba2e6c947-utilities\") pod \"redhat-marketplace-79r4z\" (UID: \"0691f390-5c98-4c03-9839-e72ba2e6c947\") " pod="openshift-marketplace/redhat-marketplace-79r4z" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.558272 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0b6ae2d-bfaf-4535-9906-31b19d4fe489-secret-volume\") pod \"c0b6ae2d-bfaf-4535-9906-31b19d4fe489\" (UID: \"c0b6ae2d-bfaf-4535-9906-31b19d4fe489\") " Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.558339 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgd64\" (UniqueName: \"kubernetes.io/projected/c0b6ae2d-bfaf-4535-9906-31b19d4fe489-kube-api-access-bgd64\") pod \"c0b6ae2d-bfaf-4535-9906-31b19d4fe489\" (UID: \"c0b6ae2d-bfaf-4535-9906-31b19d4fe489\") " Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.558406 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0b6ae2d-bfaf-4535-9906-31b19d4fe489-config-volume\") pod \"c0b6ae2d-bfaf-4535-9906-31b19d4fe489\" (UID: \"c0b6ae2d-bfaf-4535-9906-31b19d4fe489\") " Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.558644 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0691f390-5c98-4c03-9839-e72ba2e6c947-catalog-content\") pod \"redhat-marketplace-79r4z\" (UID: \"0691f390-5c98-4c03-9839-e72ba2e6c947\") " pod="openshift-marketplace/redhat-marketplace-79r4z" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.558672 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjb9v\" (UniqueName: \"kubernetes.io/projected/0691f390-5c98-4c03-9839-e72ba2e6c947-kube-api-access-rjb9v\") pod \"redhat-marketplace-79r4z\" (UID: \"0691f390-5c98-4c03-9839-e72ba2e6c947\") " pod="openshift-marketplace/redhat-marketplace-79r4z" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.558731 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0691f390-5c98-4c03-9839-e72ba2e6c947-utilities\") pod \"redhat-marketplace-79r4z\" (UID: \"0691f390-5c98-4c03-9839-e72ba2e6c947\") " pod="openshift-marketplace/redhat-marketplace-79r4z" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.560212 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0691f390-5c98-4c03-9839-e72ba2e6c947-catalog-content\") pod \"redhat-marketplace-79r4z\" (UID: \"0691f390-5c98-4c03-9839-e72ba2e6c947\") " pod="openshift-marketplace/redhat-marketplace-79r4z" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.560783 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0b6ae2d-bfaf-4535-9906-31b19d4fe489-config-volume" (OuterVolumeSpecName: "config-volume") pod "c0b6ae2d-bfaf-4535-9906-31b19d4fe489" (UID: "c0b6ae2d-bfaf-4535-9906-31b19d4fe489"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.561198 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0691f390-5c98-4c03-9839-e72ba2e6c947-utilities\") pod \"redhat-marketplace-79r4z\" (UID: \"0691f390-5c98-4c03-9839-e72ba2e6c947\") " pod="openshift-marketplace/redhat-marketplace-79r4z" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.571604 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0b6ae2d-bfaf-4535-9906-31b19d4fe489-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c0b6ae2d-bfaf-4535-9906-31b19d4fe489" (UID: "c0b6ae2d-bfaf-4535-9906-31b19d4fe489"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.571765 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0b6ae2d-bfaf-4535-9906-31b19d4fe489-kube-api-access-bgd64" (OuterVolumeSpecName: "kube-api-access-bgd64") pod "c0b6ae2d-bfaf-4535-9906-31b19d4fe489" (UID: "c0b6ae2d-bfaf-4535-9906-31b19d4fe489"). InnerVolumeSpecName "kube-api-access-bgd64". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.583329 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjb9v\" (UniqueName: \"kubernetes.io/projected/0691f390-5c98-4c03-9839-e72ba2e6c947-kube-api-access-rjb9v\") pod \"redhat-marketplace-79r4z\" (UID: \"0691f390-5c98-4c03-9839-e72ba2e6c947\") " pod="openshift-marketplace/redhat-marketplace-79r4z" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.597773 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5z654"] Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.628274 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 21 17:58:42 crc kubenswrapper[4808]: E0121 17:58:42.628535 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0b6ae2d-bfaf-4535-9906-31b19d4fe489" containerName="collect-profiles" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.628552 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0b6ae2d-bfaf-4535-9906-31b19d4fe489" containerName="collect-profiles" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.628669 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0b6ae2d-bfaf-4535-9906-31b19d4fe489" containerName="collect-profiles" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.629090 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.632709 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.632795 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.647979 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.653154 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-79r4z" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.661005 4808 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0b6ae2d-bfaf-4535-9906-31b19d4fe489-config-volume\") on node \"crc\" DevicePath \"\"" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.661043 4808 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0b6ae2d-bfaf-4535-9906-31b19d4fe489-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.661077 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgd64\" (UniqueName: \"kubernetes.io/projected/c0b6ae2d-bfaf-4535-9906-31b19d4fe489-kube-api-access-bgd64\") on node \"crc\" DevicePath \"\"" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.767157 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0c73dff9-c55e-4997-82e7-69d77858b1b7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0c73dff9-c55e-4997-82e7-69d77858b1b7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.767555 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0c73dff9-c55e-4997-82e7-69d77858b1b7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0c73dff9-c55e-4997-82e7-69d77858b1b7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.868742 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0c73dff9-c55e-4997-82e7-69d77858b1b7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0c73dff9-c55e-4997-82e7-69d77858b1b7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.868792 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0c73dff9-c55e-4997-82e7-69d77858b1b7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0c73dff9-c55e-4997-82e7-69d77858b1b7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.869452 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0c73dff9-c55e-4997-82e7-69d77858b1b7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0c73dff9-c55e-4997-82e7-69d77858b1b7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.930381 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0c73dff9-c55e-4997-82e7-69d77858b1b7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0c73dff9-c55e-4997-82e7-69d77858b1b7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 17:58:42 crc kubenswrapper[4808]: I0121 17:58:42.965403 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.054080 4808 generic.go:334] "Generic (PLEG): container finished" podID="595fab48-bf4b-443f-9367-d5718a4672e3" containerID="9de8c4953894df3916b9486ba3c379257e7ee9f1e9bbd088216a9a0099745ac6" exitCode=0 Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.054730 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5z654" event={"ID":"595fab48-bf4b-443f-9367-d5718a4672e3","Type":"ContainerDied","Data":"9de8c4953894df3916b9486ba3c379257e7ee9f1e9bbd088216a9a0099745ac6"} Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.054810 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5z654" event={"ID":"595fab48-bf4b-443f-9367-d5718a4672e3","Type":"ContainerStarted","Data":"4ad16a2763e43212346dc694520556133d7bfa3dc603edb88035c5e57e5c1398"} Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.077593 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.079357 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2" event={"ID":"c0b6ae2d-bfaf-4535-9906-31b19d4fe489","Type":"ContainerDied","Data":"a265a7647c478ac03358f72e533f89555b59cf46ff052c2a97d976b34c3410bd"} Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.079394 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a265a7647c478ac03358f72e533f89555b59cf46ff052c2a97d976b34c3410bd" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.079429 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7wlcv"] Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.080611 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7wlcv" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.087373 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.124184 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7wlcv"] Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.179052 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33777903-22e8-4520-a301-53aa0fbc9ef2-utilities\") pod \"redhat-operators-7wlcv\" (UID: \"33777903-22e8-4520-a301-53aa0fbc9ef2\") " pod="openshift-marketplace/redhat-operators-7wlcv" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.179159 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7zss\" (UniqueName: \"kubernetes.io/projected/33777903-22e8-4520-a301-53aa0fbc9ef2-kube-api-access-r7zss\") pod \"redhat-operators-7wlcv\" (UID: \"33777903-22e8-4520-a301-53aa0fbc9ef2\") " pod="openshift-marketplace/redhat-operators-7wlcv" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.179286 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33777903-22e8-4520-a301-53aa0fbc9ef2-catalog-content\") pod \"redhat-operators-7wlcv\" (UID: \"33777903-22e8-4520-a301-53aa0fbc9ef2\") " pod="openshift-marketplace/redhat-operators-7wlcv" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.226424 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-79r4z"] Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.281511 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7zss\" (UniqueName: \"kubernetes.io/projected/33777903-22e8-4520-a301-53aa0fbc9ef2-kube-api-access-r7zss\") pod \"redhat-operators-7wlcv\" (UID: \"33777903-22e8-4520-a301-53aa0fbc9ef2\") " pod="openshift-marketplace/redhat-operators-7wlcv" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.281622 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33777903-22e8-4520-a301-53aa0fbc9ef2-catalog-content\") pod \"redhat-operators-7wlcv\" (UID: \"33777903-22e8-4520-a301-53aa0fbc9ef2\") " pod="openshift-marketplace/redhat-operators-7wlcv" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.281738 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33777903-22e8-4520-a301-53aa0fbc9ef2-utilities\") pod \"redhat-operators-7wlcv\" (UID: \"33777903-22e8-4520-a301-53aa0fbc9ef2\") " pod="openshift-marketplace/redhat-operators-7wlcv" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.282208 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33777903-22e8-4520-a301-53aa0fbc9ef2-catalog-content\") pod \"redhat-operators-7wlcv\" (UID: \"33777903-22e8-4520-a301-53aa0fbc9ef2\") " pod="openshift-marketplace/redhat-operators-7wlcv" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.283666 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33777903-22e8-4520-a301-53aa0fbc9ef2-utilities\") pod \"redhat-operators-7wlcv\" (UID: \"33777903-22e8-4520-a301-53aa0fbc9ef2\") " pod="openshift-marketplace/redhat-operators-7wlcv" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.318843 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7zss\" (UniqueName: \"kubernetes.io/projected/33777903-22e8-4520-a301-53aa0fbc9ef2-kube-api-access-r7zss\") pod \"redhat-operators-7wlcv\" (UID: \"33777903-22e8-4520-a301-53aa0fbc9ef2\") " pod="openshift-marketplace/redhat-operators-7wlcv" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.325208 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.422335 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7wlcv" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.491092 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jdpdl"] Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.493062 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jdpdl" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.498215 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jdpdl"] Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.592758 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6-catalog-content\") pod \"redhat-operators-jdpdl\" (UID: \"f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6\") " pod="openshift-marketplace/redhat-operators-jdpdl" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.593391 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhj8j\" (UniqueName: \"kubernetes.io/projected/f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6-kube-api-access-fhj8j\") pod \"redhat-operators-jdpdl\" (UID: \"f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6\") " pod="openshift-marketplace/redhat-operators-jdpdl" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.593422 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6-utilities\") pod \"redhat-operators-jdpdl\" (UID: \"f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6\") " pod="openshift-marketplace/redhat-operators-jdpdl" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.695144 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6-catalog-content\") pod \"redhat-operators-jdpdl\" (UID: \"f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6\") " pod="openshift-marketplace/redhat-operators-jdpdl" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.695491 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhj8j\" (UniqueName: \"kubernetes.io/projected/f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6-kube-api-access-fhj8j\") pod \"redhat-operators-jdpdl\" (UID: \"f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6\") " pod="openshift-marketplace/redhat-operators-jdpdl" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.695512 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6-utilities\") pod \"redhat-operators-jdpdl\" (UID: \"f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6\") " pod="openshift-marketplace/redhat-operators-jdpdl" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.696097 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6-utilities\") pod \"redhat-operators-jdpdl\" (UID: \"f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6\") " pod="openshift-marketplace/redhat-operators-jdpdl" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.696328 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6-catalog-content\") pod \"redhat-operators-jdpdl\" (UID: \"f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6\") " pod="openshift-marketplace/redhat-operators-jdpdl" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.720089 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhj8j\" (UniqueName: \"kubernetes.io/projected/f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6-kube-api-access-fhj8j\") pod \"redhat-operators-jdpdl\" (UID: \"f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6\") " pod="openshift-marketplace/redhat-operators-jdpdl" Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.744197 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7wlcv"] Jan 21 17:58:43 crc kubenswrapper[4808]: W0121 17:58:43.772860 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33777903_22e8_4520_a301_53aa0fbc9ef2.slice/crio-7e69469f544bead1148dd283bddb9d4769bda103dee8d1f8fb5a0eddfa8b979d WatchSource:0}: Error finding container 7e69469f544bead1148dd283bddb9d4769bda103dee8d1f8fb5a0eddfa8b979d: Status 404 returned error can't find the container with id 7e69469f544bead1148dd283bddb9d4769bda103dee8d1f8fb5a0eddfa8b979d Jan 21 17:58:43 crc kubenswrapper[4808]: I0121 17:58:43.827416 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jdpdl" Jan 21 17:58:44 crc kubenswrapper[4808]: I0121 17:58:44.101017 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79r4z" event={"ID":"0691f390-5c98-4c03-9839-e72ba2e6c947","Type":"ContainerStarted","Data":"5268301c4d84f88fc2fca5fd9a5111338717f22a6b5908ca66abc98c3be54ec7"} Jan 21 17:58:44 crc kubenswrapper[4808]: I0121 17:58:44.125023 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0c73dff9-c55e-4997-82e7-69d77858b1b7","Type":"ContainerStarted","Data":"b868901725aae625cb63bfaccfc7e4b317c3796a621b7c37316484e224d6eed1"} Jan 21 17:58:44 crc kubenswrapper[4808]: I0121 17:58:44.128180 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7wlcv" event={"ID":"33777903-22e8-4520-a301-53aa0fbc9ef2","Type":"ContainerStarted","Data":"7e69469f544bead1148dd283bddb9d4769bda103dee8d1f8fb5a0eddfa8b979d"} Jan 21 17:58:44 crc kubenswrapper[4808]: I0121 17:58:44.522438 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jdpdl"] Jan 21 17:58:44 crc kubenswrapper[4808]: W0121 17:58:44.577376 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf53dff6f_bb26_4c73_b0be_5a7ad54f6cb6.slice/crio-d50f734f9698fdeac0f9c10ff2d9923763b952bf7feaf382c3f84c71a3c4ee68 WatchSource:0}: Error finding container d50f734f9698fdeac0f9c10ff2d9923763b952bf7feaf382c3f84c71a3c4ee68: Status 404 returned error can't find the container with id d50f734f9698fdeac0f9c10ff2d9923763b952bf7feaf382c3f84c71a3c4ee68 Jan 21 17:58:45 crc kubenswrapper[4808]: I0121 17:58:45.137820 4808 generic.go:334] "Generic (PLEG): container finished" podID="33777903-22e8-4520-a301-53aa0fbc9ef2" containerID="29df6527d44100973e0c8c31d811b6c7902efd577921ffc46fb4cb8f2812694d" exitCode=0 Jan 21 17:58:45 crc kubenswrapper[4808]: I0121 17:58:45.138184 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7wlcv" event={"ID":"33777903-22e8-4520-a301-53aa0fbc9ef2","Type":"ContainerDied","Data":"29df6527d44100973e0c8c31d811b6c7902efd577921ffc46fb4cb8f2812694d"} Jan 21 17:58:45 crc kubenswrapper[4808]: I0121 17:58:45.141337 4808 generic.go:334] "Generic (PLEG): container finished" podID="0691f390-5c98-4c03-9839-e72ba2e6c947" containerID="80ffabddaa3f7488c74769be311fc5a95bc4479e0947d9dbfb757d1c0589cbd2" exitCode=0 Jan 21 17:58:45 crc kubenswrapper[4808]: I0121 17:58:45.141405 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79r4z" event={"ID":"0691f390-5c98-4c03-9839-e72ba2e6c947","Type":"ContainerDied","Data":"80ffabddaa3f7488c74769be311fc5a95bc4479e0947d9dbfb757d1c0589cbd2"} Jan 21 17:58:45 crc kubenswrapper[4808]: I0121 17:58:45.144799 4808 generic.go:334] "Generic (PLEG): container finished" podID="0c73dff9-c55e-4997-82e7-69d77858b1b7" containerID="51fd94e0e82cfd2b71beb8f21a3692b46ea7e30f9d6e9ddedacc24d7d1e40907" exitCode=0 Jan 21 17:58:45 crc kubenswrapper[4808]: I0121 17:58:45.144870 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0c73dff9-c55e-4997-82e7-69d77858b1b7","Type":"ContainerDied","Data":"51fd94e0e82cfd2b71beb8f21a3692b46ea7e30f9d6e9ddedacc24d7d1e40907"} Jan 21 17:58:45 crc kubenswrapper[4808]: I0121 17:58:45.147024 4808 generic.go:334] "Generic (PLEG): container finished" podID="f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6" containerID="61bc74b4c9a7b0890598e054c11076b36a23c5835b24148a62486c337c8d9360" exitCode=0 Jan 21 17:58:45 crc kubenswrapper[4808]: I0121 17:58:45.147071 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdpdl" event={"ID":"f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6","Type":"ContainerDied","Data":"61bc74b4c9a7b0890598e054c11076b36a23c5835b24148a62486c337c8d9360"} Jan 21 17:58:45 crc kubenswrapper[4808]: I0121 17:58:45.147110 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdpdl" event={"ID":"f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6","Type":"ContainerStarted","Data":"d50f734f9698fdeac0f9c10ff2d9923763b952bf7feaf382c3f84c71a3c4ee68"} Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.544033 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.592380 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 21 17:58:46 crc kubenswrapper[4808]: E0121 17:58:46.592891 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c73dff9-c55e-4997-82e7-69d77858b1b7" containerName="pruner" Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.592981 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c73dff9-c55e-4997-82e7-69d77858b1b7" containerName="pruner" Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.593203 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c73dff9-c55e-4997-82e7-69d77858b1b7" containerName="pruner" Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.593760 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.596162 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.596496 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.616145 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.660198 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0c73dff9-c55e-4997-82e7-69d77858b1b7-kube-api-access\") pod \"0c73dff9-c55e-4997-82e7-69d77858b1b7\" (UID: \"0c73dff9-c55e-4997-82e7-69d77858b1b7\") " Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.660374 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0c73dff9-c55e-4997-82e7-69d77858b1b7-kubelet-dir\") pod \"0c73dff9-c55e-4997-82e7-69d77858b1b7\" (UID: \"0c73dff9-c55e-4997-82e7-69d77858b1b7\") " Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.660565 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f7f885a1-5467-4365-bcf3-2d78deba878a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f7f885a1-5467-4365-bcf3-2d78deba878a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.660699 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7f885a1-5467-4365-bcf3-2d78deba878a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f7f885a1-5467-4365-bcf3-2d78deba878a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.661791 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0c73dff9-c55e-4997-82e7-69d77858b1b7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0c73dff9-c55e-4997-82e7-69d77858b1b7" (UID: "0c73dff9-c55e-4997-82e7-69d77858b1b7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.685320 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c73dff9-c55e-4997-82e7-69d77858b1b7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0c73dff9-c55e-4997-82e7-69d77858b1b7" (UID: "0c73dff9-c55e-4997-82e7-69d77858b1b7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.761703 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7f885a1-5467-4365-bcf3-2d78deba878a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f7f885a1-5467-4365-bcf3-2d78deba878a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.761784 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f7f885a1-5467-4365-bcf3-2d78deba878a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f7f885a1-5467-4365-bcf3-2d78deba878a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.761906 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f7f885a1-5467-4365-bcf3-2d78deba878a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f7f885a1-5467-4365-bcf3-2d78deba878a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.761926 4808 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0c73dff9-c55e-4997-82e7-69d77858b1b7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.761943 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0c73dff9-c55e-4997-82e7-69d77858b1b7-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.780407 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7f885a1-5467-4365-bcf3-2d78deba878a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f7f885a1-5467-4365-bcf3-2d78deba878a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 17:58:46 crc kubenswrapper[4808]: I0121 17:58:46.922521 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 17:58:47 crc kubenswrapper[4808]: I0121 17:58:47.171742 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0c73dff9-c55e-4997-82e7-69d77858b1b7","Type":"ContainerDied","Data":"b868901725aae625cb63bfaccfc7e4b317c3796a621b7c37316484e224d6eed1"} Jan 21 17:58:47 crc kubenswrapper[4808]: I0121 17:58:47.171798 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b868901725aae625cb63bfaccfc7e4b317c3796a621b7c37316484e224d6eed1" Jan 21 17:58:47 crc kubenswrapper[4808]: I0121 17:58:47.171853 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 17:58:47 crc kubenswrapper[4808]: I0121 17:58:47.227684 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 21 17:58:48 crc kubenswrapper[4808]: I0121 17:58:48.159752 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-zjcb6" Jan 21 17:58:48 crc kubenswrapper[4808]: I0121 17:58:48.210075 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f7f885a1-5467-4365-bcf3-2d78deba878a","Type":"ContainerStarted","Data":"e96c6c48231e0b44e1bb46187654d72e6874aee2a10872ef36fcf491cc8f9437"} Jan 21 17:58:48 crc kubenswrapper[4808]: I0121 17:58:48.210127 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f7f885a1-5467-4365-bcf3-2d78deba878a","Type":"ContainerStarted","Data":"df8e8f0778d622b97127f429b2f0e44272f7eeaccf811497001773f289d3bcb1"} Jan 21 17:58:48 crc kubenswrapper[4808]: I0121 17:58:48.238070 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.23795961 podStartE2EDuration="2.23795961s" podCreationTimestamp="2026-01-21 17:58:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:58:48.233057418 +0000 UTC m=+162.963568313" watchObservedRunningTime="2026-01-21 17:58:48.23795961 +0000 UTC m=+162.968470495" Jan 21 17:58:49 crc kubenswrapper[4808]: I0121 17:58:49.254931 4808 generic.go:334] "Generic (PLEG): container finished" podID="f7f885a1-5467-4365-bcf3-2d78deba878a" containerID="e96c6c48231e0b44e1bb46187654d72e6874aee2a10872ef36fcf491cc8f9437" exitCode=0 Jan 21 17:58:49 crc kubenswrapper[4808]: I0121 17:58:49.254993 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f7f885a1-5467-4365-bcf3-2d78deba878a","Type":"ContainerDied","Data":"e96c6c48231e0b44e1bb46187654d72e6874aee2a10872ef36fcf491cc8f9437"} Jan 21 17:58:51 crc kubenswrapper[4808]: I0121 17:58:51.441197 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs\") pod \"network-metrics-daemon-gm2t2\" (UID: \"8d39c565-f404-451e-896f-ecc5af76a000\") " pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:58:51 crc kubenswrapper[4808]: I0121 17:58:51.449365 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8d39c565-f404-451e-896f-ecc5af76a000-metrics-certs\") pod \"network-metrics-daemon-gm2t2\" (UID: \"8d39c565-f404-451e-896f-ecc5af76a000\") " pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:58:51 crc kubenswrapper[4808]: I0121 17:58:51.560867 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gm2t2" Jan 21 17:58:52 crc kubenswrapper[4808]: I0121 17:58:52.000657 4808 patch_prober.go:28] interesting pod/console-f9d7485db-h6gm7 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Jan 21 17:58:52 crc kubenswrapper[4808]: I0121 17:58:52.000712 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-h6gm7" podUID="05688b08-2c9f-4c6c-967a-ff64de7837c0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Jan 21 17:58:52 crc kubenswrapper[4808]: I0121 17:58:52.261758 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-ftpvl" Jan 21 17:58:56 crc kubenswrapper[4808]: I0121 17:58:56.144681 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 17:58:56 crc kubenswrapper[4808]: I0121 17:58:56.212960 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7f885a1-5467-4365-bcf3-2d78deba878a-kube-api-access\") pod \"f7f885a1-5467-4365-bcf3-2d78deba878a\" (UID: \"f7f885a1-5467-4365-bcf3-2d78deba878a\") " Jan 21 17:58:56 crc kubenswrapper[4808]: I0121 17:58:56.213012 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f7f885a1-5467-4365-bcf3-2d78deba878a-kubelet-dir\") pod \"f7f885a1-5467-4365-bcf3-2d78deba878a\" (UID: \"f7f885a1-5467-4365-bcf3-2d78deba878a\") " Jan 21 17:58:56 crc kubenswrapper[4808]: I0121 17:58:56.213218 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7f885a1-5467-4365-bcf3-2d78deba878a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f7f885a1-5467-4365-bcf3-2d78deba878a" (UID: "f7f885a1-5467-4365-bcf3-2d78deba878a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 17:58:56 crc kubenswrapper[4808]: I0121 17:58:56.213392 4808 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f7f885a1-5467-4365-bcf3-2d78deba878a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 21 17:58:56 crc kubenswrapper[4808]: I0121 17:58:56.232067 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7f885a1-5467-4365-bcf3-2d78deba878a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f7f885a1-5467-4365-bcf3-2d78deba878a" (UID: "f7f885a1-5467-4365-bcf3-2d78deba878a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:58:56 crc kubenswrapper[4808]: I0121 17:58:56.309279 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f7f885a1-5467-4365-bcf3-2d78deba878a","Type":"ContainerDied","Data":"df8e8f0778d622b97127f429b2f0e44272f7eeaccf811497001773f289d3bcb1"} Jan 21 17:58:56 crc kubenswrapper[4808]: I0121 17:58:56.309329 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df8e8f0778d622b97127f429b2f0e44272f7eeaccf811497001773f289d3bcb1" Jan 21 17:58:56 crc kubenswrapper[4808]: I0121 17:58:56.309374 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 17:58:56 crc kubenswrapper[4808]: I0121 17:58:56.322430 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7f885a1-5467-4365-bcf3-2d78deba878a-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 17:58:59 crc kubenswrapper[4808]: I0121 17:58:59.599640 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 17:58:59 crc kubenswrapper[4808]: I0121 17:58:59.599991 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 17:59:00 crc kubenswrapper[4808]: I0121 17:59:00.498936 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 17:59:02 crc kubenswrapper[4808]: I0121 17:59:02.100059 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:59:02 crc kubenswrapper[4808]: I0121 17:59:02.105221 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-h6gm7" Jan 21 17:59:11 crc kubenswrapper[4808]: I0121 17:59:11.819826 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 17:59:13 crc kubenswrapper[4808]: I0121 17:59:13.062550 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q8wfq" Jan 21 17:59:20 crc kubenswrapper[4808]: I0121 17:59:20.593521 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 21 17:59:20 crc kubenswrapper[4808]: E0121 17:59:20.594719 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7f885a1-5467-4365-bcf3-2d78deba878a" containerName="pruner" Jan 21 17:59:20 crc kubenswrapper[4808]: I0121 17:59:20.594744 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7f885a1-5467-4365-bcf3-2d78deba878a" containerName="pruner" Jan 21 17:59:20 crc kubenswrapper[4808]: I0121 17:59:20.594941 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7f885a1-5467-4365-bcf3-2d78deba878a" containerName="pruner" Jan 21 17:59:20 crc kubenswrapper[4808]: I0121 17:59:20.596220 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 17:59:20 crc kubenswrapper[4808]: I0121 17:59:20.598877 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 21 17:59:20 crc kubenswrapper[4808]: I0121 17:59:20.598928 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 21 17:59:20 crc kubenswrapper[4808]: I0121 17:59:20.604230 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 21 17:59:20 crc kubenswrapper[4808]: I0121 17:59:20.718468 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a9f4a034-5fe4-4c43-b723-abddd8f0d1af-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a9f4a034-5fe4-4c43-b723-abddd8f0d1af\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 17:59:20 crc kubenswrapper[4808]: I0121 17:59:20.718606 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a9f4a034-5fe4-4c43-b723-abddd8f0d1af-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a9f4a034-5fe4-4c43-b723-abddd8f0d1af\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 17:59:20 crc kubenswrapper[4808]: I0121 17:59:20.820943 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a9f4a034-5fe4-4c43-b723-abddd8f0d1af-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a9f4a034-5fe4-4c43-b723-abddd8f0d1af\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 17:59:20 crc kubenswrapper[4808]: I0121 17:59:20.821129 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a9f4a034-5fe4-4c43-b723-abddd8f0d1af-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a9f4a034-5fe4-4c43-b723-abddd8f0d1af\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 17:59:20 crc kubenswrapper[4808]: I0121 17:59:20.821232 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a9f4a034-5fe4-4c43-b723-abddd8f0d1af-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a9f4a034-5fe4-4c43-b723-abddd8f0d1af\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 17:59:20 crc kubenswrapper[4808]: I0121 17:59:20.840467 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a9f4a034-5fe4-4c43-b723-abddd8f0d1af-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a9f4a034-5fe4-4c43-b723-abddd8f0d1af\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 17:59:20 crc kubenswrapper[4808]: I0121 17:59:20.917014 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 17:59:21 crc kubenswrapper[4808]: E0121 17:59:21.988816 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 21 17:59:21 crc kubenswrapper[4808]: E0121 17:59:21.989332 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7xzbc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-nr5dg_openshift-marketplace(6221711c-1051-44e9-8fad-029c73d0ae71): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 21 17:59:21 crc kubenswrapper[4808]: E0121 17:59:21.992092 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-nr5dg" podUID="6221711c-1051-44e9-8fad-029c73d0ae71" Jan 21 17:59:22 crc kubenswrapper[4808]: E0121 17:59:22.077592 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 21 17:59:22 crc kubenswrapper[4808]: E0121 17:59:22.077780 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mztfc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-hfz2d_openshift-marketplace(09939aec-8ef7-4b4d-a934-9f00e19bc7ed): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 21 17:59:22 crc kubenswrapper[4808]: E0121 17:59:22.079051 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-hfz2d" podUID="09939aec-8ef7-4b4d-a934-9f00e19bc7ed" Jan 21 17:59:23 crc kubenswrapper[4808]: E0121 17:59:23.306785 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-nr5dg" podUID="6221711c-1051-44e9-8fad-029c73d0ae71" Jan 21 17:59:23 crc kubenswrapper[4808]: E0121 17:59:23.307335 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-hfz2d" podUID="09939aec-8ef7-4b4d-a934-9f00e19bc7ed" Jan 21 17:59:23 crc kubenswrapper[4808]: E0121 17:59:23.375608 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 21 17:59:23 crc kubenswrapper[4808]: E0121 17:59:23.375832 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-clswl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-5z654_openshift-marketplace(595fab48-bf4b-443f-9367-d5718a4672e3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 21 17:59:23 crc kubenswrapper[4808]: E0121 17:59:23.377875 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-5z654" podUID="595fab48-bf4b-443f-9367-d5718a4672e3" Jan 21 17:59:24 crc kubenswrapper[4808]: E0121 17:59:24.380551 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-5z654" podUID="595fab48-bf4b-443f-9367-d5718a4672e3" Jan 21 17:59:24 crc kubenswrapper[4808]: E0121 17:59:24.392801 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 21 17:59:24 crc kubenswrapper[4808]: E0121 17:59:24.392980 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rjb9v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-79r4z_openshift-marketplace(0691f390-5c98-4c03-9839-e72ba2e6c947): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 21 17:59:24 crc kubenswrapper[4808]: E0121 17:59:24.394185 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-79r4z" podUID="0691f390-5c98-4c03-9839-e72ba2e6c947" Jan 21 17:59:24 crc kubenswrapper[4808]: E0121 17:59:24.454575 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 21 17:59:24 crc kubenswrapper[4808]: E0121 17:59:24.454763 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m6f5w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-2lzcv_openshift-marketplace(6bb3e28c-45c6-4992-beee-69dfd1ae7d55): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 21 17:59:24 crc kubenswrapper[4808]: E0121 17:59:24.455927 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-2lzcv" podUID="6bb3e28c-45c6-4992-beee-69dfd1ae7d55" Jan 21 17:59:24 crc kubenswrapper[4808]: E0121 17:59:24.467956 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 21 17:59:24 crc kubenswrapper[4808]: E0121 17:59:24.468196 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6tggj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-4fsv9_openshift-marketplace(3a84f07d-c20d-4a1d-a272-118d29bcdb60): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 21 17:59:24 crc kubenswrapper[4808]: E0121 17:59:24.469649 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-4fsv9" podUID="3a84f07d-c20d-4a1d-a272-118d29bcdb60" Jan 21 17:59:25 crc kubenswrapper[4808]: I0121 17:59:25.597237 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 21 17:59:25 crc kubenswrapper[4808]: I0121 17:59:25.602802 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 21 17:59:25 crc kubenswrapper[4808]: I0121 17:59:25.602991 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 21 17:59:25 crc kubenswrapper[4808]: I0121 17:59:25.693075 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f152312-7210-4e7a-b9c1-6a2942f488aa-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2f152312-7210-4e7a-b9c1-6a2942f488aa\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 17:59:25 crc kubenswrapper[4808]: I0121 17:59:25.693121 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f152312-7210-4e7a-b9c1-6a2942f488aa-kube-api-access\") pod \"installer-9-crc\" (UID: \"2f152312-7210-4e7a-b9c1-6a2942f488aa\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 17:59:25 crc kubenswrapper[4808]: I0121 17:59:25.693153 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2f152312-7210-4e7a-b9c1-6a2942f488aa-var-lock\") pod \"installer-9-crc\" (UID: \"2f152312-7210-4e7a-b9c1-6a2942f488aa\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 17:59:25 crc kubenswrapper[4808]: I0121 17:59:25.793802 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f152312-7210-4e7a-b9c1-6a2942f488aa-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2f152312-7210-4e7a-b9c1-6a2942f488aa\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 17:59:25 crc kubenswrapper[4808]: I0121 17:59:25.793869 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f152312-7210-4e7a-b9c1-6a2942f488aa-kube-api-access\") pod \"installer-9-crc\" (UID: \"2f152312-7210-4e7a-b9c1-6a2942f488aa\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 17:59:25 crc kubenswrapper[4808]: I0121 17:59:25.793912 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2f152312-7210-4e7a-b9c1-6a2942f488aa-var-lock\") pod \"installer-9-crc\" (UID: \"2f152312-7210-4e7a-b9c1-6a2942f488aa\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 17:59:25 crc kubenswrapper[4808]: I0121 17:59:25.794010 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f152312-7210-4e7a-b9c1-6a2942f488aa-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2f152312-7210-4e7a-b9c1-6a2942f488aa\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 17:59:25 crc kubenswrapper[4808]: I0121 17:59:25.794036 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2f152312-7210-4e7a-b9c1-6a2942f488aa-var-lock\") pod \"installer-9-crc\" (UID: \"2f152312-7210-4e7a-b9c1-6a2942f488aa\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 17:59:25 crc kubenswrapper[4808]: I0121 17:59:25.819212 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f152312-7210-4e7a-b9c1-6a2942f488aa-kube-api-access\") pod \"installer-9-crc\" (UID: \"2f152312-7210-4e7a-b9c1-6a2942f488aa\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 17:59:25 crc kubenswrapper[4808]: I0121 17:59:25.934386 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 21 17:59:28 crc kubenswrapper[4808]: E0121 17:59:28.321069 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-2lzcv" podUID="6bb3e28c-45c6-4992-beee-69dfd1ae7d55" Jan 21 17:59:28 crc kubenswrapper[4808]: E0121 17:59:28.321120 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-79r4z" podUID="0691f390-5c98-4c03-9839-e72ba2e6c947" Jan 21 17:59:28 crc kubenswrapper[4808]: E0121 17:59:28.321226 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-4fsv9" podUID="3a84f07d-c20d-4a1d-a272-118d29bcdb60" Jan 21 17:59:28 crc kubenswrapper[4808]: E0121 17:59:28.355946 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 21 17:59:28 crc kubenswrapper[4808]: E0121 17:59:28.356118 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r7zss,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-7wlcv_openshift-marketplace(33777903-22e8-4520-a301-53aa0fbc9ef2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 21 17:59:28 crc kubenswrapper[4808]: E0121 17:59:28.357824 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-7wlcv" podUID="33777903-22e8-4520-a301-53aa0fbc9ef2" Jan 21 17:59:28 crc kubenswrapper[4808]: E0121 17:59:28.374702 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 21 17:59:28 crc kubenswrapper[4808]: E0121 17:59:28.374870 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fhj8j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-jdpdl_openshift-marketplace(f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 21 17:59:28 crc kubenswrapper[4808]: E0121 17:59:28.376408 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-jdpdl" podUID="f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6" Jan 21 17:59:28 crc kubenswrapper[4808]: E0121 17:59:28.490268 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-7wlcv" podUID="33777903-22e8-4520-a301-53aa0fbc9ef2" Jan 21 17:59:28 crc kubenswrapper[4808]: I0121 17:59:28.656967 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 21 17:59:28 crc kubenswrapper[4808]: I0121 17:59:28.782150 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 21 17:59:28 crc kubenswrapper[4808]: I0121 17:59:28.795551 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-gm2t2"] Jan 21 17:59:28 crc kubenswrapper[4808]: W0121 17:59:28.805447 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda9f4a034_5fe4_4c43_b723_abddd8f0d1af.slice/crio-d45be9e8bbbec25e253790e8e769bb95a8d1cf6df8e715e2c4d8417d8c9ae9e8 WatchSource:0}: Error finding container d45be9e8bbbec25e253790e8e769bb95a8d1cf6df8e715e2c4d8417d8c9ae9e8: Status 404 returned error can't find the container with id d45be9e8bbbec25e253790e8e769bb95a8d1cf6df8e715e2c4d8417d8c9ae9e8 Jan 21 17:59:28 crc kubenswrapper[4808]: W0121 17:59:28.808333 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d39c565_f404_451e_896f_ecc5af76a000.slice/crio-a28bb99a85df513f9a75f35e8d6071af8fa4b2dab6d29abf8efc3c0c2202fcb6 WatchSource:0}: Error finding container a28bb99a85df513f9a75f35e8d6071af8fa4b2dab6d29abf8efc3c0c2202fcb6: Status 404 returned error can't find the container with id a28bb99a85df513f9a75f35e8d6071af8fa4b2dab6d29abf8efc3c0c2202fcb6 Jan 21 17:59:29 crc kubenswrapper[4808]: I0121 17:59:29.513596 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a9f4a034-5fe4-4c43-b723-abddd8f0d1af","Type":"ContainerStarted","Data":"d45be9e8bbbec25e253790e8e769bb95a8d1cf6df8e715e2c4d8417d8c9ae9e8"} Jan 21 17:59:29 crc kubenswrapper[4808]: I0121 17:59:29.515544 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gm2t2" event={"ID":"8d39c565-f404-451e-896f-ecc5af76a000","Type":"ContainerStarted","Data":"a28bb99a85df513f9a75f35e8d6071af8fa4b2dab6d29abf8efc3c0c2202fcb6"} Jan 21 17:59:29 crc kubenswrapper[4808]: I0121 17:59:29.517146 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2f152312-7210-4e7a-b9c1-6a2942f488aa","Type":"ContainerStarted","Data":"39cd1f403cdd0ab5909b5d3b32782969ee12f74e7ffbb3231c202f91b04dadae"} Jan 21 17:59:29 crc kubenswrapper[4808]: I0121 17:59:29.604873 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 17:59:29 crc kubenswrapper[4808]: I0121 17:59:29.604968 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 17:59:30 crc kubenswrapper[4808]: I0121 17:59:30.524315 4808 generic.go:334] "Generic (PLEG): container finished" podID="a9f4a034-5fe4-4c43-b723-abddd8f0d1af" containerID="7167e759d8690e4f14b720c5e24392ff814bf71c40b56e2b18091f5e4d70dbbc" exitCode=0 Jan 21 17:59:30 crc kubenswrapper[4808]: I0121 17:59:30.524435 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a9f4a034-5fe4-4c43-b723-abddd8f0d1af","Type":"ContainerDied","Data":"7167e759d8690e4f14b720c5e24392ff814bf71c40b56e2b18091f5e4d70dbbc"} Jan 21 17:59:30 crc kubenswrapper[4808]: I0121 17:59:30.527801 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gm2t2" event={"ID":"8d39c565-f404-451e-896f-ecc5af76a000","Type":"ContainerStarted","Data":"e31bd7eeabeaf3fe5b1646f2150325387d0d50e0de6cc10f4425f8df4f412bf5"} Jan 21 17:59:30 crc kubenswrapper[4808]: I0121 17:59:30.527853 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gm2t2" event={"ID":"8d39c565-f404-451e-896f-ecc5af76a000","Type":"ContainerStarted","Data":"eb9a7c4004d5921dbef536a2ab565001dca1e89b6387fb41144b27be9fb4f5dd"} Jan 21 17:59:30 crc kubenswrapper[4808]: I0121 17:59:30.533149 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2f152312-7210-4e7a-b9c1-6a2942f488aa","Type":"ContainerStarted","Data":"474fc1c7bd09c01331f20a863ed7d0de5601ad2931cd0abd3dffba99dc627033"} Jan 21 17:59:30 crc kubenswrapper[4808]: I0121 17:59:30.578768 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=5.578745383 podStartE2EDuration="5.578745383s" podCreationTimestamp="2026-01-21 17:59:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:59:30.555830545 +0000 UTC m=+205.286341430" watchObservedRunningTime="2026-01-21 17:59:30.578745383 +0000 UTC m=+205.309256278" Jan 21 17:59:31 crc kubenswrapper[4808]: I0121 17:59:31.800284 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 17:59:31 crc kubenswrapper[4808]: I0121 17:59:31.816933 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-gm2t2" podStartSLOduration=182.816911082 podStartE2EDuration="3m2.816911082s" podCreationTimestamp="2026-01-21 17:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 17:59:30.57703675 +0000 UTC m=+205.307547645" watchObservedRunningTime="2026-01-21 17:59:31.816911082 +0000 UTC m=+206.547421967" Jan 21 17:59:31 crc kubenswrapper[4808]: I0121 17:59:31.975504 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a9f4a034-5fe4-4c43-b723-abddd8f0d1af-kubelet-dir\") pod \"a9f4a034-5fe4-4c43-b723-abddd8f0d1af\" (UID: \"a9f4a034-5fe4-4c43-b723-abddd8f0d1af\") " Jan 21 17:59:31 crc kubenswrapper[4808]: I0121 17:59:31.975611 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a9f4a034-5fe4-4c43-b723-abddd8f0d1af-kube-api-access\") pod \"a9f4a034-5fe4-4c43-b723-abddd8f0d1af\" (UID: \"a9f4a034-5fe4-4c43-b723-abddd8f0d1af\") " Jan 21 17:59:31 crc kubenswrapper[4808]: I0121 17:59:31.975737 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9f4a034-5fe4-4c43-b723-abddd8f0d1af-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a9f4a034-5fe4-4c43-b723-abddd8f0d1af" (UID: "a9f4a034-5fe4-4c43-b723-abddd8f0d1af"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 17:59:31 crc kubenswrapper[4808]: I0121 17:59:31.975879 4808 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a9f4a034-5fe4-4c43-b723-abddd8f0d1af-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 21 17:59:31 crc kubenswrapper[4808]: I0121 17:59:31.982739 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9f4a034-5fe4-4c43-b723-abddd8f0d1af-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a9f4a034-5fe4-4c43-b723-abddd8f0d1af" (UID: "a9f4a034-5fe4-4c43-b723-abddd8f0d1af"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 17:59:32 crc kubenswrapper[4808]: I0121 17:59:32.076863 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a9f4a034-5fe4-4c43-b723-abddd8f0d1af-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 17:59:32 crc kubenswrapper[4808]: I0121 17:59:32.546161 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a9f4a034-5fe4-4c43-b723-abddd8f0d1af","Type":"ContainerDied","Data":"d45be9e8bbbec25e253790e8e769bb95a8d1cf6df8e715e2c4d8417d8c9ae9e8"} Jan 21 17:59:32 crc kubenswrapper[4808]: I0121 17:59:32.546214 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d45be9e8bbbec25e253790e8e769bb95a8d1cf6df8e715e2c4d8417d8c9ae9e8" Jan 21 17:59:32 crc kubenswrapper[4808]: I0121 17:59:32.546303 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 17:59:37 crc kubenswrapper[4808]: I0121 17:59:37.569876 4808 generic.go:334] "Generic (PLEG): container finished" podID="09939aec-8ef7-4b4d-a934-9f00e19bc7ed" containerID="5d2782118c5a9caeccd1ad931941f28491a03fac09d79d81cd649ed9fa951f36" exitCode=0 Jan 21 17:59:37 crc kubenswrapper[4808]: I0121 17:59:37.570336 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfz2d" event={"ID":"09939aec-8ef7-4b4d-a934-9f00e19bc7ed","Type":"ContainerDied","Data":"5d2782118c5a9caeccd1ad931941f28491a03fac09d79d81cd649ed9fa951f36"} Jan 21 17:59:38 crc kubenswrapper[4808]: I0121 17:59:38.578720 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfz2d" event={"ID":"09939aec-8ef7-4b4d-a934-9f00e19bc7ed","Type":"ContainerStarted","Data":"7437495b13c0ee5355d0872d6d2b0f60265acfe6d82f2aef48c0f302015d9b21"} Jan 21 17:59:38 crc kubenswrapper[4808]: I0121 17:59:38.600278 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hfz2d" podStartSLOduration=2.629708832 podStartE2EDuration="59.600261409s" podCreationTimestamp="2026-01-21 17:58:39 +0000 UTC" firstStartedPulling="2026-01-21 17:58:40.999982809 +0000 UTC m=+155.730493694" lastFinishedPulling="2026-01-21 17:59:37.970535346 +0000 UTC m=+212.701046271" observedRunningTime="2026-01-21 17:59:38.59832275 +0000 UTC m=+213.328833655" watchObservedRunningTime="2026-01-21 17:59:38.600261409 +0000 UTC m=+213.330772294" Jan 21 17:59:39 crc kubenswrapper[4808]: I0121 17:59:39.585072 4808 generic.go:334] "Generic (PLEG): container finished" podID="6221711c-1051-44e9-8fad-029c73d0ae71" containerID="d8efaad2463d18200c90985bee0cd483174d57b2dc9f5ac387a3e9a4334521a9" exitCode=0 Jan 21 17:59:39 crc kubenswrapper[4808]: I0121 17:59:39.585160 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nr5dg" event={"ID":"6221711c-1051-44e9-8fad-029c73d0ae71","Type":"ContainerDied","Data":"d8efaad2463d18200c90985bee0cd483174d57b2dc9f5ac387a3e9a4334521a9"} Jan 21 17:59:40 crc kubenswrapper[4808]: I0121 17:59:40.198583 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hfz2d" Jan 21 17:59:40 crc kubenswrapper[4808]: I0121 17:59:40.198949 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hfz2d" Jan 21 17:59:40 crc kubenswrapper[4808]: I0121 17:59:40.252838 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hfz2d" Jan 21 17:59:40 crc kubenswrapper[4808]: I0121 17:59:40.601143 4808 generic.go:334] "Generic (PLEG): container finished" podID="595fab48-bf4b-443f-9367-d5718a4672e3" containerID="4426bba94623b11c0458d424883aebca4109ae1076ed7b8809f31ce7c2ae3346" exitCode=0 Jan 21 17:59:40 crc kubenswrapper[4808]: I0121 17:59:40.601218 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5z654" event={"ID":"595fab48-bf4b-443f-9367-d5718a4672e3","Type":"ContainerDied","Data":"4426bba94623b11c0458d424883aebca4109ae1076ed7b8809f31ce7c2ae3346"} Jan 21 17:59:40 crc kubenswrapper[4808]: I0121 17:59:40.607569 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nr5dg" event={"ID":"6221711c-1051-44e9-8fad-029c73d0ae71","Type":"ContainerStarted","Data":"8ae75eab40906e64e505ae531c2f1a82a5ce7af7b08c3ba54c31a0a6dc6f4e23"} Jan 21 17:59:40 crc kubenswrapper[4808]: I0121 17:59:40.644024 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nr5dg" podStartSLOduration=2.665649402 podStartE2EDuration="1m0.644001779s" podCreationTimestamp="2026-01-21 17:58:40 +0000 UTC" firstStartedPulling="2026-01-21 17:58:42.018095139 +0000 UTC m=+156.748606034" lastFinishedPulling="2026-01-21 17:59:39.996447526 +0000 UTC m=+214.726958411" observedRunningTime="2026-01-21 17:59:40.637774042 +0000 UTC m=+215.368284927" watchObservedRunningTime="2026-01-21 17:59:40.644001779 +0000 UTC m=+215.374512674" Jan 21 17:59:40 crc kubenswrapper[4808]: I0121 17:59:40.674209 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nr5dg" Jan 21 17:59:40 crc kubenswrapper[4808]: I0121 17:59:40.674284 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nr5dg" Jan 21 17:59:41 crc kubenswrapper[4808]: I0121 17:59:41.614449 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5z654" event={"ID":"595fab48-bf4b-443f-9367-d5718a4672e3","Type":"ContainerStarted","Data":"31f8a0c477d427da1d65d61c40baa83d5f8bd7d10fd8ebe1ab14f5397b34c31c"} Jan 21 17:59:41 crc kubenswrapper[4808]: I0121 17:59:41.634754 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5z654" podStartSLOduration=2.680218683 podStartE2EDuration="1m0.634736152s" podCreationTimestamp="2026-01-21 17:58:41 +0000 UTC" firstStartedPulling="2026-01-21 17:58:43.056220116 +0000 UTC m=+157.786731011" lastFinishedPulling="2026-01-21 17:59:41.010737595 +0000 UTC m=+215.741248480" observedRunningTime="2026-01-21 17:59:41.631627574 +0000 UTC m=+216.362138509" watchObservedRunningTime="2026-01-21 17:59:41.634736152 +0000 UTC m=+216.365247037" Jan 21 17:59:41 crc kubenswrapper[4808]: I0121 17:59:41.712778 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-nr5dg" podUID="6221711c-1051-44e9-8fad-029c73d0ae71" containerName="registry-server" probeResult="failure" output=< Jan 21 17:59:41 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Jan 21 17:59:41 crc kubenswrapper[4808]: > Jan 21 17:59:42 crc kubenswrapper[4808]: I0121 17:59:42.194402 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5z654" Jan 21 17:59:42 crc kubenswrapper[4808]: I0121 17:59:42.194439 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5z654" Jan 21 17:59:43 crc kubenswrapper[4808]: I0121 17:59:43.243273 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-5z654" podUID="595fab48-bf4b-443f-9367-d5718a4672e3" containerName="registry-server" probeResult="failure" output=< Jan 21 17:59:43 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Jan 21 17:59:43 crc kubenswrapper[4808]: > Jan 21 17:59:50 crc kubenswrapper[4808]: I0121 17:59:50.248847 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hfz2d" Jan 21 17:59:50 crc kubenswrapper[4808]: I0121 17:59:50.714604 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nr5dg" Jan 21 17:59:50 crc kubenswrapper[4808]: I0121 17:59:50.760226 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nr5dg" Jan 21 17:59:51 crc kubenswrapper[4808]: I0121 17:59:51.478950 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nr5dg"] Jan 21 17:59:52 crc kubenswrapper[4808]: I0121 17:59:52.231816 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5z654" Jan 21 17:59:52 crc kubenswrapper[4808]: I0121 17:59:52.281271 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5z654" Jan 21 17:59:52 crc kubenswrapper[4808]: I0121 17:59:52.670749 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nr5dg" podUID="6221711c-1051-44e9-8fad-029c73d0ae71" containerName="registry-server" containerID="cri-o://8ae75eab40906e64e505ae531c2f1a82a5ce7af7b08c3ba54c31a0a6dc6f4e23" gracePeriod=2 Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.680045 4808 generic.go:334] "Generic (PLEG): container finished" podID="6221711c-1051-44e9-8fad-029c73d0ae71" containerID="8ae75eab40906e64e505ae531c2f1a82a5ce7af7b08c3ba54c31a0a6dc6f4e23" exitCode=0 Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.680121 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nr5dg" event={"ID":"6221711c-1051-44e9-8fad-029c73d0ae71","Type":"ContainerDied","Data":"8ae75eab40906e64e505ae531c2f1a82a5ce7af7b08c3ba54c31a0a6dc6f4e23"} Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.724923 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-5v5c4"] Jan 21 17:59:53 crc kubenswrapper[4808]: E0121 17:59:53.725147 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9f4a034-5fe4-4c43-b723-abddd8f0d1af" containerName="pruner" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.725160 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9f4a034-5fe4-4c43-b723-abddd8f0d1af" containerName="pruner" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.725275 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9f4a034-5fe4-4c43-b723-abddd8f0d1af" containerName="pruner" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.725640 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.740736 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-5v5c4"] Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.865057 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/524e5263-8f59-4b6a-9920-c6c885c1f8a6-registry-tls\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.865102 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/524e5263-8f59-4b6a-9920-c6c885c1f8a6-trusted-ca\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.865138 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.865172 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/524e5263-8f59-4b6a-9920-c6c885c1f8a6-registry-certificates\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.865412 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/524e5263-8f59-4b6a-9920-c6c885c1f8a6-installation-pull-secrets\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.865576 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/524e5263-8f59-4b6a-9920-c6c885c1f8a6-bound-sa-token\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.865681 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/524e5263-8f59-4b6a-9920-c6c885c1f8a6-ca-trust-extracted\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.865781 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q44g2\" (UniqueName: \"kubernetes.io/projected/524e5263-8f59-4b6a-9920-c6c885c1f8a6-kube-api-access-q44g2\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.889673 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.966841 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/524e5263-8f59-4b6a-9920-c6c885c1f8a6-registry-tls\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.966890 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/524e5263-8f59-4b6a-9920-c6c885c1f8a6-trusted-ca\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.966931 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/524e5263-8f59-4b6a-9920-c6c885c1f8a6-registry-certificates\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.966957 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/524e5263-8f59-4b6a-9920-c6c885c1f8a6-installation-pull-secrets\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.966984 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/524e5263-8f59-4b6a-9920-c6c885c1f8a6-bound-sa-token\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.967006 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/524e5263-8f59-4b6a-9920-c6c885c1f8a6-ca-trust-extracted\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.967034 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q44g2\" (UniqueName: \"kubernetes.io/projected/524e5263-8f59-4b6a-9920-c6c885c1f8a6-kube-api-access-q44g2\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.968328 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/524e5263-8f59-4b6a-9920-c6c885c1f8a6-trusted-ca\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.968410 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/524e5263-8f59-4b6a-9920-c6c885c1f8a6-ca-trust-extracted\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.968480 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/524e5263-8f59-4b6a-9920-c6c885c1f8a6-registry-certificates\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.973133 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/524e5263-8f59-4b6a-9920-c6c885c1f8a6-installation-pull-secrets\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.973157 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/524e5263-8f59-4b6a-9920-c6c885c1f8a6-registry-tls\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:53 crc kubenswrapper[4808]: I0121 17:59:53.983877 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/524e5263-8f59-4b6a-9920-c6c885c1f8a6-bound-sa-token\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:54 crc kubenswrapper[4808]: I0121 17:59:54.000116 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q44g2\" (UniqueName: \"kubernetes.io/projected/524e5263-8f59-4b6a-9920-c6c885c1f8a6-kube-api-access-q44g2\") pod \"image-registry-66df7c8f76-5v5c4\" (UID: \"524e5263-8f59-4b6a-9920-c6c885c1f8a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:54 crc kubenswrapper[4808]: I0121 17:59:54.042976 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 17:59:59 crc kubenswrapper[4808]: I0121 17:59:59.599515 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 17:59:59 crc kubenswrapper[4808]: I0121 17:59:59.600228 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 17:59:59 crc kubenswrapper[4808]: I0121 17:59:59.600423 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 17:59:59 crc kubenswrapper[4808]: I0121 17:59:59.600993 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb"} pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 17:59:59 crc kubenswrapper[4808]: I0121 17:59:59.601124 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" containerID="cri-o://d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb" gracePeriod=600 Jan 21 18:00:00 crc kubenswrapper[4808]: I0121 18:00:00.129327 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q"] Jan 21 18:00:00 crc kubenswrapper[4808]: I0121 18:00:00.132071 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" Jan 21 18:00:00 crc kubenswrapper[4808]: I0121 18:00:00.138636 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 21 18:00:00 crc kubenswrapper[4808]: I0121 18:00:00.138653 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 21 18:00:00 crc kubenswrapper[4808]: I0121 18:00:00.140993 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q"] Jan 21 18:00:00 crc kubenswrapper[4808]: I0121 18:00:00.270629 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ae9be85d-1887-4d2b-95ad-a71cdd7f8f10-secret-volume\") pod \"collect-profiles-29483640-bx86q\" (UID: \"ae9be85d-1887-4d2b-95ad-a71cdd7f8f10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" Jan 21 18:00:00 crc kubenswrapper[4808]: I0121 18:00:00.270683 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29zqb\" (UniqueName: \"kubernetes.io/projected/ae9be85d-1887-4d2b-95ad-a71cdd7f8f10-kube-api-access-29zqb\") pod \"collect-profiles-29483640-bx86q\" (UID: \"ae9be85d-1887-4d2b-95ad-a71cdd7f8f10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" Jan 21 18:00:00 crc kubenswrapper[4808]: I0121 18:00:00.271026 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae9be85d-1887-4d2b-95ad-a71cdd7f8f10-config-volume\") pod \"collect-profiles-29483640-bx86q\" (UID: \"ae9be85d-1887-4d2b-95ad-a71cdd7f8f10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" Jan 21 18:00:00 crc kubenswrapper[4808]: I0121 18:00:00.372603 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae9be85d-1887-4d2b-95ad-a71cdd7f8f10-config-volume\") pod \"collect-profiles-29483640-bx86q\" (UID: \"ae9be85d-1887-4d2b-95ad-a71cdd7f8f10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" Jan 21 18:00:00 crc kubenswrapper[4808]: I0121 18:00:00.372997 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ae9be85d-1887-4d2b-95ad-a71cdd7f8f10-secret-volume\") pod \"collect-profiles-29483640-bx86q\" (UID: \"ae9be85d-1887-4d2b-95ad-a71cdd7f8f10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" Jan 21 18:00:00 crc kubenswrapper[4808]: I0121 18:00:00.373024 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29zqb\" (UniqueName: \"kubernetes.io/projected/ae9be85d-1887-4d2b-95ad-a71cdd7f8f10-kube-api-access-29zqb\") pod \"collect-profiles-29483640-bx86q\" (UID: \"ae9be85d-1887-4d2b-95ad-a71cdd7f8f10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" Jan 21 18:00:00 crc kubenswrapper[4808]: I0121 18:00:00.374471 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae9be85d-1887-4d2b-95ad-a71cdd7f8f10-config-volume\") pod \"collect-profiles-29483640-bx86q\" (UID: \"ae9be85d-1887-4d2b-95ad-a71cdd7f8f10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" Jan 21 18:00:00 crc kubenswrapper[4808]: I0121 18:00:00.379221 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ae9be85d-1887-4d2b-95ad-a71cdd7f8f10-secret-volume\") pod \"collect-profiles-29483640-bx86q\" (UID: \"ae9be85d-1887-4d2b-95ad-a71cdd7f8f10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" Jan 21 18:00:00 crc kubenswrapper[4808]: I0121 18:00:00.389928 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29zqb\" (UniqueName: \"kubernetes.io/projected/ae9be85d-1887-4d2b-95ad-a71cdd7f8f10-kube-api-access-29zqb\") pod \"collect-profiles-29483640-bx86q\" (UID: \"ae9be85d-1887-4d2b-95ad-a71cdd7f8f10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" Jan 21 18:00:00 crc kubenswrapper[4808]: I0121 18:00:00.449270 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" Jan 21 18:00:00 crc kubenswrapper[4808]: E0121 18:00:00.674598 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8ae75eab40906e64e505ae531c2f1a82a5ce7af7b08c3ba54c31a0a6dc6f4e23 is running failed: container process not found" containerID="8ae75eab40906e64e505ae531c2f1a82a5ce7af7b08c3ba54c31a0a6dc6f4e23" cmd=["grpc_health_probe","-addr=:50051"] Jan 21 18:00:00 crc kubenswrapper[4808]: E0121 18:00:00.675474 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8ae75eab40906e64e505ae531c2f1a82a5ce7af7b08c3ba54c31a0a6dc6f4e23 is running failed: container process not found" containerID="8ae75eab40906e64e505ae531c2f1a82a5ce7af7b08c3ba54c31a0a6dc6f4e23" cmd=["grpc_health_probe","-addr=:50051"] Jan 21 18:00:00 crc kubenswrapper[4808]: E0121 18:00:00.675903 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8ae75eab40906e64e505ae531c2f1a82a5ce7af7b08c3ba54c31a0a6dc6f4e23 is running failed: container process not found" containerID="8ae75eab40906e64e505ae531c2f1a82a5ce7af7b08c3ba54c31a0a6dc6f4e23" cmd=["grpc_health_probe","-addr=:50051"] Jan 21 18:00:00 crc kubenswrapper[4808]: E0121 18:00:00.675942 4808 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8ae75eab40906e64e505ae531c2f1a82a5ce7af7b08c3ba54c31a0a6dc6f4e23 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-nr5dg" podUID="6221711c-1051-44e9-8fad-029c73d0ae71" containerName="registry-server" Jan 21 18:00:00 crc kubenswrapper[4808]: I0121 18:00:00.714941 4808 generic.go:334] "Generic (PLEG): container finished" podID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerID="d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb" exitCode=0 Jan 21 18:00:00 crc kubenswrapper[4808]: I0121 18:00:00.715030 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerDied","Data":"d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb"} Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.322968 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2lzcv"] Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.331207 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4fsv9"] Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.338775 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hfz2d"] Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.339081 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hfz2d" podUID="09939aec-8ef7-4b4d-a934-9f00e19bc7ed" containerName="registry-server" containerID="cri-o://7437495b13c0ee5355d0872d6d2b0f60265acfe6d82f2aef48c0f302015d9b21" gracePeriod=30 Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.352894 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pvk7n"] Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.353145 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" podUID="5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5" containerName="marketplace-operator" containerID="cri-o://cbd58dac49caa9a8cf3d13700bf902392cd9d8cbe2e1cb48e8dbe6e67b4cf935" gracePeriod=30 Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.367640 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cklmp"] Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.368555 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cklmp" Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.371300 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5z654"] Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.371642 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5z654" podUID="595fab48-bf4b-443f-9367-d5718a4672e3" containerName="registry-server" containerID="cri-o://31f8a0c477d427da1d65d61c40baa83d5f8bd7d10fd8ebe1ab14f5397b34c31c" gracePeriod=30 Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.380732 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-79r4z"] Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.387214 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cklmp"] Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.391150 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7wlcv"] Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.394297 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jdpdl"] Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.487045 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0a6a5330-18dc-4ab0-bfb0-6b98549e28bc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cklmp\" (UID: \"0a6a5330-18dc-4ab0-bfb0-6b98549e28bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-cklmp" Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.487293 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0a6a5330-18dc-4ab0-bfb0-6b98549e28bc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cklmp\" (UID: \"0a6a5330-18dc-4ab0-bfb0-6b98549e28bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-cklmp" Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.487452 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7twxb\" (UniqueName: \"kubernetes.io/projected/0a6a5330-18dc-4ab0-bfb0-6b98549e28bc-kube-api-access-7twxb\") pod \"marketplace-operator-79b997595-cklmp\" (UID: \"0a6a5330-18dc-4ab0-bfb0-6b98549e28bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-cklmp" Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.588910 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7twxb\" (UniqueName: \"kubernetes.io/projected/0a6a5330-18dc-4ab0-bfb0-6b98549e28bc-kube-api-access-7twxb\") pod \"marketplace-operator-79b997595-cklmp\" (UID: \"0a6a5330-18dc-4ab0-bfb0-6b98549e28bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-cklmp" Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.589001 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0a6a5330-18dc-4ab0-bfb0-6b98549e28bc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cklmp\" (UID: \"0a6a5330-18dc-4ab0-bfb0-6b98549e28bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-cklmp" Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.589076 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0a6a5330-18dc-4ab0-bfb0-6b98549e28bc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cklmp\" (UID: \"0a6a5330-18dc-4ab0-bfb0-6b98549e28bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-cklmp" Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.591703 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0a6a5330-18dc-4ab0-bfb0-6b98549e28bc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cklmp\" (UID: \"0a6a5330-18dc-4ab0-bfb0-6b98549e28bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-cklmp" Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.606098 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0a6a5330-18dc-4ab0-bfb0-6b98549e28bc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cklmp\" (UID: \"0a6a5330-18dc-4ab0-bfb0-6b98549e28bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-cklmp" Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.612221 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7twxb\" (UniqueName: \"kubernetes.io/projected/0a6a5330-18dc-4ab0-bfb0-6b98549e28bc-kube-api-access-7twxb\") pod \"marketplace-operator-79b997595-cklmp\" (UID: \"0a6a5330-18dc-4ab0-bfb0-6b98549e28bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-cklmp" Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.689099 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cklmp" Jan 21 18:00:01 crc kubenswrapper[4808]: I0121 18:00:01.734566 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bz45s"] Jan 21 18:00:02 crc kubenswrapper[4808]: E0121 18:00:02.194125 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31f8a0c477d427da1d65d61c40baa83d5f8bd7d10fd8ebe1ab14f5397b34c31c is running failed: container process not found" containerID="31f8a0c477d427da1d65d61c40baa83d5f8bd7d10fd8ebe1ab14f5397b34c31c" cmd=["grpc_health_probe","-addr=:50051"] Jan 21 18:00:02 crc kubenswrapper[4808]: E0121 18:00:02.195140 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31f8a0c477d427da1d65d61c40baa83d5f8bd7d10fd8ebe1ab14f5397b34c31c is running failed: container process not found" containerID="31f8a0c477d427da1d65d61c40baa83d5f8bd7d10fd8ebe1ab14f5397b34c31c" cmd=["grpc_health_probe","-addr=:50051"] Jan 21 18:00:02 crc kubenswrapper[4808]: E0121 18:00:02.195728 4808 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31f8a0c477d427da1d65d61c40baa83d5f8bd7d10fd8ebe1ab14f5397b34c31c is running failed: container process not found" containerID="31f8a0c477d427da1d65d61c40baa83d5f8bd7d10fd8ebe1ab14f5397b34c31c" cmd=["grpc_health_probe","-addr=:50051"] Jan 21 18:00:02 crc kubenswrapper[4808]: E0121 18:00:02.195833 4808 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31f8a0c477d427da1d65d61c40baa83d5f8bd7d10fd8ebe1ab14f5397b34c31c is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-5z654" podUID="595fab48-bf4b-443f-9367-d5718a4672e3" containerName="registry-server" Jan 21 18:00:02 crc kubenswrapper[4808]: I0121 18:00:02.738405 4808 generic.go:334] "Generic (PLEG): container finished" podID="5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5" containerID="cbd58dac49caa9a8cf3d13700bf902392cd9d8cbe2e1cb48e8dbe6e67b4cf935" exitCode=0 Jan 21 18:00:02 crc kubenswrapper[4808]: I0121 18:00:02.738458 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" event={"ID":"5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5","Type":"ContainerDied","Data":"cbd58dac49caa9a8cf3d13700bf902392cd9d8cbe2e1cb48e8dbe6e67b4cf935"} Jan 21 18:00:02 crc kubenswrapper[4808]: I0121 18:00:02.943868 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nr5dg" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.005835 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xzbc\" (UniqueName: \"kubernetes.io/projected/6221711c-1051-44e9-8fad-029c73d0ae71-kube-api-access-7xzbc\") pod \"6221711c-1051-44e9-8fad-029c73d0ae71\" (UID: \"6221711c-1051-44e9-8fad-029c73d0ae71\") " Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.005914 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6221711c-1051-44e9-8fad-029c73d0ae71-utilities\") pod \"6221711c-1051-44e9-8fad-029c73d0ae71\" (UID: \"6221711c-1051-44e9-8fad-029c73d0ae71\") " Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.005984 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6221711c-1051-44e9-8fad-029c73d0ae71-catalog-content\") pod \"6221711c-1051-44e9-8fad-029c73d0ae71\" (UID: \"6221711c-1051-44e9-8fad-029c73d0ae71\") " Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.006829 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6221711c-1051-44e9-8fad-029c73d0ae71-utilities" (OuterVolumeSpecName: "utilities") pod "6221711c-1051-44e9-8fad-029c73d0ae71" (UID: "6221711c-1051-44e9-8fad-029c73d0ae71"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.011268 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6221711c-1051-44e9-8fad-029c73d0ae71-kube-api-access-7xzbc" (OuterVolumeSpecName: "kube-api-access-7xzbc") pod "6221711c-1051-44e9-8fad-029c73d0ae71" (UID: "6221711c-1051-44e9-8fad-029c73d0ae71"). InnerVolumeSpecName "kube-api-access-7xzbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.057583 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6221711c-1051-44e9-8fad-029c73d0ae71-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6221711c-1051-44e9-8fad-029c73d0ae71" (UID: "6221711c-1051-44e9-8fad-029c73d0ae71"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.122553 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6221711c-1051-44e9-8fad-029c73d0ae71-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.122612 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6221711c-1051-44e9-8fad-029c73d0ae71-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.122626 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xzbc\" (UniqueName: \"kubernetes.io/projected/6221711c-1051-44e9-8fad-029c73d0ae71-kube-api-access-7xzbc\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.147904 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.193458 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hfz2d" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.195889 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5z654" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.223526 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09939aec-8ef7-4b4d-a934-9f00e19bc7ed-utilities\") pod \"09939aec-8ef7-4b4d-a934-9f00e19bc7ed\" (UID: \"09939aec-8ef7-4b4d-a934-9f00e19bc7ed\") " Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.223586 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09939aec-8ef7-4b4d-a934-9f00e19bc7ed-catalog-content\") pod \"09939aec-8ef7-4b4d-a934-9f00e19bc7ed\" (UID: \"09939aec-8ef7-4b4d-a934-9f00e19bc7ed\") " Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.223630 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mztfc\" (UniqueName: \"kubernetes.io/projected/09939aec-8ef7-4b4d-a934-9f00e19bc7ed-kube-api-access-mztfc\") pod \"09939aec-8ef7-4b4d-a934-9f00e19bc7ed\" (UID: \"09939aec-8ef7-4b4d-a934-9f00e19bc7ed\") " Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.223648 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/595fab48-bf4b-443f-9367-d5718a4672e3-catalog-content\") pod \"595fab48-bf4b-443f-9367-d5718a4672e3\" (UID: \"595fab48-bf4b-443f-9367-d5718a4672e3\") " Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.223706 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5-marketplace-trusted-ca\") pod \"5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5\" (UID: \"5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5\") " Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.223727 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/595fab48-bf4b-443f-9367-d5718a4672e3-utilities\") pod \"595fab48-bf4b-443f-9367-d5718a4672e3\" (UID: \"595fab48-bf4b-443f-9367-d5718a4672e3\") " Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.223798 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5-marketplace-operator-metrics\") pod \"5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5\" (UID: \"5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5\") " Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.223831 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clswl\" (UniqueName: \"kubernetes.io/projected/595fab48-bf4b-443f-9367-d5718a4672e3-kube-api-access-clswl\") pod \"595fab48-bf4b-443f-9367-d5718a4672e3\" (UID: \"595fab48-bf4b-443f-9367-d5718a4672e3\") " Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.223868 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7prvs\" (UniqueName: \"kubernetes.io/projected/5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5-kube-api-access-7prvs\") pod \"5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5\" (UID: \"5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5\") " Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.224597 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09939aec-8ef7-4b4d-a934-9f00e19bc7ed-utilities" (OuterVolumeSpecName: "utilities") pod "09939aec-8ef7-4b4d-a934-9f00e19bc7ed" (UID: "09939aec-8ef7-4b4d-a934-9f00e19bc7ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.224722 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5" (UID: "5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.225738 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/595fab48-bf4b-443f-9367-d5718a4672e3-utilities" (OuterVolumeSpecName: "utilities") pod "595fab48-bf4b-443f-9367-d5718a4672e3" (UID: "595fab48-bf4b-443f-9367-d5718a4672e3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.227689 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5-kube-api-access-7prvs" (OuterVolumeSpecName: "kube-api-access-7prvs") pod "5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5" (UID: "5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5"). InnerVolumeSpecName "kube-api-access-7prvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.228096 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09939aec-8ef7-4b4d-a934-9f00e19bc7ed-kube-api-access-mztfc" (OuterVolumeSpecName: "kube-api-access-mztfc") pod "09939aec-8ef7-4b4d-a934-9f00e19bc7ed" (UID: "09939aec-8ef7-4b4d-a934-9f00e19bc7ed"). InnerVolumeSpecName "kube-api-access-mztfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.228881 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5" (UID: "5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.234683 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/595fab48-bf4b-443f-9367-d5718a4672e3-kube-api-access-clswl" (OuterVolumeSpecName: "kube-api-access-clswl") pod "595fab48-bf4b-443f-9367-d5718a4672e3" (UID: "595fab48-bf4b-443f-9367-d5718a4672e3"). InnerVolumeSpecName "kube-api-access-clswl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.259530 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/595fab48-bf4b-443f-9367-d5718a4672e3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "595fab48-bf4b-443f-9367-d5718a4672e3" (UID: "595fab48-bf4b-443f-9367-d5718a4672e3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.290373 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09939aec-8ef7-4b4d-a934-9f00e19bc7ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09939aec-8ef7-4b4d-a934-9f00e19bc7ed" (UID: "09939aec-8ef7-4b4d-a934-9f00e19bc7ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.326133 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7prvs\" (UniqueName: \"kubernetes.io/projected/5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5-kube-api-access-7prvs\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.326180 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09939aec-8ef7-4b4d-a934-9f00e19bc7ed-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.326195 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09939aec-8ef7-4b4d-a934-9f00e19bc7ed-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.326208 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mztfc\" (UniqueName: \"kubernetes.io/projected/09939aec-8ef7-4b4d-a934-9f00e19bc7ed-kube-api-access-mztfc\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.326221 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/595fab48-bf4b-443f-9367-d5718a4672e3-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.326236 4808 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.326267 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/595fab48-bf4b-443f-9367-d5718a4672e3-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.326278 4808 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.326290 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clswl\" (UniqueName: \"kubernetes.io/projected/595fab48-bf4b-443f-9367-d5718a4672e3-kube-api-access-clswl\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.396385 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-5v5c4"] Jan 21 18:00:03 crc kubenswrapper[4808]: W0121 18:00:03.439137 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod524e5263_8f59_4b6a_9920_c6c885c1f8a6.slice/crio-f44a36ac4672b62b8eb64a9494478935a4a14b3db8161a59ea80f419f9efcfbd WatchSource:0}: Error finding container f44a36ac4672b62b8eb64a9494478935a4a14b3db8161a59ea80f419f9efcfbd: Status 404 returned error can't find the container with id f44a36ac4672b62b8eb64a9494478935a4a14b3db8161a59ea80f419f9efcfbd Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.571993 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cklmp"] Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.629645 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q"] Jan 21 18:00:03 crc kubenswrapper[4808]: W0121 18:00:03.648688 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae9be85d_1887_4d2b_95ad_a71cdd7f8f10.slice/crio-72eb31190dd59a8354ced5449537c7bd199bc87e0a99ebb8bad4d14ef977414b WatchSource:0}: Error finding container 72eb31190dd59a8354ced5449537c7bd199bc87e0a99ebb8bad4d14ef977414b: Status 404 returned error can't find the container with id 72eb31190dd59a8354ced5449537c7bd199bc87e0a99ebb8bad4d14ef977414b Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.745698 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nr5dg" event={"ID":"6221711c-1051-44e9-8fad-029c73d0ae71","Type":"ContainerDied","Data":"f1b07c4835ad0926217706d2ddf4498515cc5ff0bf5cd300ef2894226ce471f7"} Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.745753 4808 scope.go:117] "RemoveContainer" containerID="8ae75eab40906e64e505ae531c2f1a82a5ce7af7b08c3ba54c31a0a6dc6f4e23" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.745880 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nr5dg" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.751331 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerStarted","Data":"b1f94a9dbae93a06cda5ec2fa475694e62e99f13f0c15187ffea4d991182c486"} Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.756809 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.756855 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" event={"ID":"5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5","Type":"ContainerDied","Data":"7d9f6eddf2262022cc1b232550ad1e5a53b71fd2cb73860ce8c91672eedcf75b"} Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.760137 4808 generic.go:334] "Generic (PLEG): container finished" podID="09939aec-8ef7-4b4d-a934-9f00e19bc7ed" containerID="7437495b13c0ee5355d0872d6d2b0f60265acfe6d82f2aef48c0f302015d9b21" exitCode=0 Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.760181 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hfz2d" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.760204 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfz2d" event={"ID":"09939aec-8ef7-4b4d-a934-9f00e19bc7ed","Type":"ContainerDied","Data":"7437495b13c0ee5355d0872d6d2b0f60265acfe6d82f2aef48c0f302015d9b21"} Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.760233 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfz2d" event={"ID":"09939aec-8ef7-4b4d-a934-9f00e19bc7ed","Type":"ContainerDied","Data":"5cdc40ff058c140c7aa79d59e5c3b7bdaea72112b20f4d6beb57b31bb9f1dd07"} Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.761552 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" event={"ID":"ae9be85d-1887-4d2b-95ad-a71cdd7f8f10","Type":"ContainerStarted","Data":"72eb31190dd59a8354ced5449537c7bd199bc87e0a99ebb8bad4d14ef977414b"} Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.764882 4808 generic.go:334] "Generic (PLEG): container finished" podID="595fab48-bf4b-443f-9367-d5718a4672e3" containerID="31f8a0c477d427da1d65d61c40baa83d5f8bd7d10fd8ebe1ab14f5397b34c31c" exitCode=0 Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.764913 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5z654" event={"ID":"595fab48-bf4b-443f-9367-d5718a4672e3","Type":"ContainerDied","Data":"31f8a0c477d427da1d65d61c40baa83d5f8bd7d10fd8ebe1ab14f5397b34c31c"} Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.764954 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5z654" event={"ID":"595fab48-bf4b-443f-9367-d5718a4672e3","Type":"ContainerDied","Data":"4ad16a2763e43212346dc694520556133d7bfa3dc603edb88035c5e57e5c1398"} Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.766042 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cklmp" event={"ID":"0a6a5330-18dc-4ab0-bfb0-6b98549e28bc","Type":"ContainerStarted","Data":"29017390c0b866420b5c5c90598b1adc2c6b162c080d519b7ad67484208807e9"} Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.766344 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5z654" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.767253 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" event={"ID":"524e5263-8f59-4b6a-9920-c6c885c1f8a6","Type":"ContainerStarted","Data":"f44a36ac4672b62b8eb64a9494478935a4a14b3db8161a59ea80f419f9efcfbd"} Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.773665 4808 scope.go:117] "RemoveContainer" containerID="d8efaad2463d18200c90985bee0cd483174d57b2dc9f5ac387a3e9a4334521a9" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.792794 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nr5dg"] Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.793830 4808 scope.go:117] "RemoveContainer" containerID="97ffdcaab65867b1dd6ac8390bb30fc9d2cc128c836e8f1e0bfd561a53ff6f0a" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.796294 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nr5dg"] Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.803730 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pvk7n"] Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.813802 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pvk7n"] Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.825974 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5z654"] Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.833203 4808 scope.go:117] "RemoveContainer" containerID="cbd58dac49caa9a8cf3d13700bf902392cd9d8cbe2e1cb48e8dbe6e67b4cf935" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.846395 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5z654"] Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.857102 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hfz2d"] Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.861719 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hfz2d"] Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.865989 4808 scope.go:117] "RemoveContainer" containerID="7437495b13c0ee5355d0872d6d2b0f60265acfe6d82f2aef48c0f302015d9b21" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.922355 4808 scope.go:117] "RemoveContainer" containerID="5d2782118c5a9caeccd1ad931941f28491a03fac09d79d81cd649ed9fa951f36" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.936278 4808 scope.go:117] "RemoveContainer" containerID="d25da3ea931b69f3f101613565c2596843abdaa238cd7b071f8a804bb697956c" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.957153 4808 scope.go:117] "RemoveContainer" containerID="7437495b13c0ee5355d0872d6d2b0f60265acfe6d82f2aef48c0f302015d9b21" Jan 21 18:00:03 crc kubenswrapper[4808]: E0121 18:00:03.957700 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7437495b13c0ee5355d0872d6d2b0f60265acfe6d82f2aef48c0f302015d9b21\": container with ID starting with 7437495b13c0ee5355d0872d6d2b0f60265acfe6d82f2aef48c0f302015d9b21 not found: ID does not exist" containerID="7437495b13c0ee5355d0872d6d2b0f60265acfe6d82f2aef48c0f302015d9b21" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.957801 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7437495b13c0ee5355d0872d6d2b0f60265acfe6d82f2aef48c0f302015d9b21"} err="failed to get container status \"7437495b13c0ee5355d0872d6d2b0f60265acfe6d82f2aef48c0f302015d9b21\": rpc error: code = NotFound desc = could not find container \"7437495b13c0ee5355d0872d6d2b0f60265acfe6d82f2aef48c0f302015d9b21\": container with ID starting with 7437495b13c0ee5355d0872d6d2b0f60265acfe6d82f2aef48c0f302015d9b21 not found: ID does not exist" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.957968 4808 scope.go:117] "RemoveContainer" containerID="5d2782118c5a9caeccd1ad931941f28491a03fac09d79d81cd649ed9fa951f36" Jan 21 18:00:03 crc kubenswrapper[4808]: E0121 18:00:03.958864 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d2782118c5a9caeccd1ad931941f28491a03fac09d79d81cd649ed9fa951f36\": container with ID starting with 5d2782118c5a9caeccd1ad931941f28491a03fac09d79d81cd649ed9fa951f36 not found: ID does not exist" containerID="5d2782118c5a9caeccd1ad931941f28491a03fac09d79d81cd649ed9fa951f36" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.958953 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d2782118c5a9caeccd1ad931941f28491a03fac09d79d81cd649ed9fa951f36"} err="failed to get container status \"5d2782118c5a9caeccd1ad931941f28491a03fac09d79d81cd649ed9fa951f36\": rpc error: code = NotFound desc = could not find container \"5d2782118c5a9caeccd1ad931941f28491a03fac09d79d81cd649ed9fa951f36\": container with ID starting with 5d2782118c5a9caeccd1ad931941f28491a03fac09d79d81cd649ed9fa951f36 not found: ID does not exist" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.959036 4808 scope.go:117] "RemoveContainer" containerID="d25da3ea931b69f3f101613565c2596843abdaa238cd7b071f8a804bb697956c" Jan 21 18:00:03 crc kubenswrapper[4808]: E0121 18:00:03.959347 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d25da3ea931b69f3f101613565c2596843abdaa238cd7b071f8a804bb697956c\": container with ID starting with d25da3ea931b69f3f101613565c2596843abdaa238cd7b071f8a804bb697956c not found: ID does not exist" containerID="d25da3ea931b69f3f101613565c2596843abdaa238cd7b071f8a804bb697956c" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.959382 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d25da3ea931b69f3f101613565c2596843abdaa238cd7b071f8a804bb697956c"} err="failed to get container status \"d25da3ea931b69f3f101613565c2596843abdaa238cd7b071f8a804bb697956c\": rpc error: code = NotFound desc = could not find container \"d25da3ea931b69f3f101613565c2596843abdaa238cd7b071f8a804bb697956c\": container with ID starting with d25da3ea931b69f3f101613565c2596843abdaa238cd7b071f8a804bb697956c not found: ID does not exist" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.959406 4808 scope.go:117] "RemoveContainer" containerID="31f8a0c477d427da1d65d61c40baa83d5f8bd7d10fd8ebe1ab14f5397b34c31c" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.973702 4808 scope.go:117] "RemoveContainer" containerID="4426bba94623b11c0458d424883aebca4109ae1076ed7b8809f31ce7c2ae3346" Jan 21 18:00:03 crc kubenswrapper[4808]: I0121 18:00:03.987769 4808 scope.go:117] "RemoveContainer" containerID="9de8c4953894df3916b9486ba3c379257e7ee9f1e9bbd088216a9a0099745ac6" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.009507 4808 scope.go:117] "RemoveContainer" containerID="31f8a0c477d427da1d65d61c40baa83d5f8bd7d10fd8ebe1ab14f5397b34c31c" Jan 21 18:00:04 crc kubenswrapper[4808]: E0121 18:00:04.010223 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31f8a0c477d427da1d65d61c40baa83d5f8bd7d10fd8ebe1ab14f5397b34c31c\": container with ID starting with 31f8a0c477d427da1d65d61c40baa83d5f8bd7d10fd8ebe1ab14f5397b34c31c not found: ID does not exist" containerID="31f8a0c477d427da1d65d61c40baa83d5f8bd7d10fd8ebe1ab14f5397b34c31c" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.010293 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31f8a0c477d427da1d65d61c40baa83d5f8bd7d10fd8ebe1ab14f5397b34c31c"} err="failed to get container status \"31f8a0c477d427da1d65d61c40baa83d5f8bd7d10fd8ebe1ab14f5397b34c31c\": rpc error: code = NotFound desc = could not find container \"31f8a0c477d427da1d65d61c40baa83d5f8bd7d10fd8ebe1ab14f5397b34c31c\": container with ID starting with 31f8a0c477d427da1d65d61c40baa83d5f8bd7d10fd8ebe1ab14f5397b34c31c not found: ID does not exist" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.010328 4808 scope.go:117] "RemoveContainer" containerID="4426bba94623b11c0458d424883aebca4109ae1076ed7b8809f31ce7c2ae3346" Jan 21 18:00:04 crc kubenswrapper[4808]: E0121 18:00:04.010775 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4426bba94623b11c0458d424883aebca4109ae1076ed7b8809f31ce7c2ae3346\": container with ID starting with 4426bba94623b11c0458d424883aebca4109ae1076ed7b8809f31ce7c2ae3346 not found: ID does not exist" containerID="4426bba94623b11c0458d424883aebca4109ae1076ed7b8809f31ce7c2ae3346" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.010815 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4426bba94623b11c0458d424883aebca4109ae1076ed7b8809f31ce7c2ae3346"} err="failed to get container status \"4426bba94623b11c0458d424883aebca4109ae1076ed7b8809f31ce7c2ae3346\": rpc error: code = NotFound desc = could not find container \"4426bba94623b11c0458d424883aebca4109ae1076ed7b8809f31ce7c2ae3346\": container with ID starting with 4426bba94623b11c0458d424883aebca4109ae1076ed7b8809f31ce7c2ae3346 not found: ID does not exist" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.010838 4808 scope.go:117] "RemoveContainer" containerID="9de8c4953894df3916b9486ba3c379257e7ee9f1e9bbd088216a9a0099745ac6" Jan 21 18:00:04 crc kubenswrapper[4808]: E0121 18:00:04.011187 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9de8c4953894df3916b9486ba3c379257e7ee9f1e9bbd088216a9a0099745ac6\": container with ID starting with 9de8c4953894df3916b9486ba3c379257e7ee9f1e9bbd088216a9a0099745ac6 not found: ID does not exist" containerID="9de8c4953894df3916b9486ba3c379257e7ee9f1e9bbd088216a9a0099745ac6" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.011215 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9de8c4953894df3916b9486ba3c379257e7ee9f1e9bbd088216a9a0099745ac6"} err="failed to get container status \"9de8c4953894df3916b9486ba3c379257e7ee9f1e9bbd088216a9a0099745ac6\": rpc error: code = NotFound desc = could not find container \"9de8c4953894df3916b9486ba3c379257e7ee9f1e9bbd088216a9a0099745ac6\": container with ID starting with 9de8c4953894df3916b9486ba3c379257e7ee9f1e9bbd088216a9a0099745ac6 not found: ID does not exist" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.121623 4808 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-pvk7n container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.121715 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-pvk7n" podUID="5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.355200 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lx72r"] Jan 21 18:00:04 crc kubenswrapper[4808]: E0121 18:00:04.355829 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5" containerName="marketplace-operator" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.356043 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5" containerName="marketplace-operator" Jan 21 18:00:04 crc kubenswrapper[4808]: E0121 18:00:04.356169 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09939aec-8ef7-4b4d-a934-9f00e19bc7ed" containerName="extract-content" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.356336 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="09939aec-8ef7-4b4d-a934-9f00e19bc7ed" containerName="extract-content" Jan 21 18:00:04 crc kubenswrapper[4808]: E0121 18:00:04.356483 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09939aec-8ef7-4b4d-a934-9f00e19bc7ed" containerName="extract-utilities" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.356596 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="09939aec-8ef7-4b4d-a934-9f00e19bc7ed" containerName="extract-utilities" Jan 21 18:00:04 crc kubenswrapper[4808]: E0121 18:00:04.356729 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09939aec-8ef7-4b4d-a934-9f00e19bc7ed" containerName="registry-server" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.356837 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="09939aec-8ef7-4b4d-a934-9f00e19bc7ed" containerName="registry-server" Jan 21 18:00:04 crc kubenswrapper[4808]: E0121 18:00:04.356945 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="595fab48-bf4b-443f-9367-d5718a4672e3" containerName="extract-content" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.357051 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="595fab48-bf4b-443f-9367-d5718a4672e3" containerName="extract-content" Jan 21 18:00:04 crc kubenswrapper[4808]: E0121 18:00:04.357176 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6221711c-1051-44e9-8fad-029c73d0ae71" containerName="extract-content" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.357329 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="6221711c-1051-44e9-8fad-029c73d0ae71" containerName="extract-content" Jan 21 18:00:04 crc kubenswrapper[4808]: E0121 18:00:04.357462 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6221711c-1051-44e9-8fad-029c73d0ae71" containerName="registry-server" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.357580 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="6221711c-1051-44e9-8fad-029c73d0ae71" containerName="registry-server" Jan 21 18:00:04 crc kubenswrapper[4808]: E0121 18:00:04.357691 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="595fab48-bf4b-443f-9367-d5718a4672e3" containerName="registry-server" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.357835 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="595fab48-bf4b-443f-9367-d5718a4672e3" containerName="registry-server" Jan 21 18:00:04 crc kubenswrapper[4808]: E0121 18:00:04.357994 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6221711c-1051-44e9-8fad-029c73d0ae71" containerName="extract-utilities" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.358135 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="6221711c-1051-44e9-8fad-029c73d0ae71" containerName="extract-utilities" Jan 21 18:00:04 crc kubenswrapper[4808]: E0121 18:00:04.358324 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="595fab48-bf4b-443f-9367-d5718a4672e3" containerName="extract-utilities" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.358456 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="595fab48-bf4b-443f-9367-d5718a4672e3" containerName="extract-utilities" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.358746 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5" containerName="marketplace-operator" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.358869 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="595fab48-bf4b-443f-9367-d5718a4672e3" containerName="registry-server" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.359109 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="09939aec-8ef7-4b4d-a934-9f00e19bc7ed" containerName="registry-server" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.359275 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="6221711c-1051-44e9-8fad-029c73d0ae71" containerName="registry-server" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.360680 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lx72r" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.363432 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lx72r"] Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.463758 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhbxc\" (UniqueName: \"kubernetes.io/projected/b05e5192-b448-446a-96ba-3ec53d3644bf-kube-api-access-nhbxc\") pod \"redhat-marketplace-lx72r\" (UID: \"b05e5192-b448-446a-96ba-3ec53d3644bf\") " pod="openshift-marketplace/redhat-marketplace-lx72r" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.463860 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b05e5192-b448-446a-96ba-3ec53d3644bf-utilities\") pod \"redhat-marketplace-lx72r\" (UID: \"b05e5192-b448-446a-96ba-3ec53d3644bf\") " pod="openshift-marketplace/redhat-marketplace-lx72r" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.463905 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b05e5192-b448-446a-96ba-3ec53d3644bf-catalog-content\") pod \"redhat-marketplace-lx72r\" (UID: \"b05e5192-b448-446a-96ba-3ec53d3644bf\") " pod="openshift-marketplace/redhat-marketplace-lx72r" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.564770 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhbxc\" (UniqueName: \"kubernetes.io/projected/b05e5192-b448-446a-96ba-3ec53d3644bf-kube-api-access-nhbxc\") pod \"redhat-marketplace-lx72r\" (UID: \"b05e5192-b448-446a-96ba-3ec53d3644bf\") " pod="openshift-marketplace/redhat-marketplace-lx72r" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.564842 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b05e5192-b448-446a-96ba-3ec53d3644bf-utilities\") pod \"redhat-marketplace-lx72r\" (UID: \"b05e5192-b448-446a-96ba-3ec53d3644bf\") " pod="openshift-marketplace/redhat-marketplace-lx72r" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.564894 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b05e5192-b448-446a-96ba-3ec53d3644bf-catalog-content\") pod \"redhat-marketplace-lx72r\" (UID: \"b05e5192-b448-446a-96ba-3ec53d3644bf\") " pod="openshift-marketplace/redhat-marketplace-lx72r" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.638080 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b05e5192-b448-446a-96ba-3ec53d3644bf-utilities\") pod \"redhat-marketplace-lx72r\" (UID: \"b05e5192-b448-446a-96ba-3ec53d3644bf\") " pod="openshift-marketplace/redhat-marketplace-lx72r" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.638197 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b05e5192-b448-446a-96ba-3ec53d3644bf-catalog-content\") pod \"redhat-marketplace-lx72r\" (UID: \"b05e5192-b448-446a-96ba-3ec53d3644bf\") " pod="openshift-marketplace/redhat-marketplace-lx72r" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.646536 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhbxc\" (UniqueName: \"kubernetes.io/projected/b05e5192-b448-446a-96ba-3ec53d3644bf-kube-api-access-nhbxc\") pod \"redhat-marketplace-lx72r\" (UID: \"b05e5192-b448-446a-96ba-3ec53d3644bf\") " pod="openshift-marketplace/redhat-marketplace-lx72r" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.688193 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lx72r" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.783971 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4fsv9" event={"ID":"3a84f07d-c20d-4a1d-a272-118d29bcdb60","Type":"ContainerStarted","Data":"917d69008a1ee33f31200c809e8c9a31224d1fdb26bd388c4c074af8d0d90ab7"} Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.784143 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4fsv9" podUID="3a84f07d-c20d-4a1d-a272-118d29bcdb60" containerName="extract-content" containerID="cri-o://917d69008a1ee33f31200c809e8c9a31224d1fdb26bd388c4c074af8d0d90ab7" gracePeriod=30 Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.790017 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" event={"ID":"524e5263-8f59-4b6a-9920-c6c885c1f8a6","Type":"ContainerStarted","Data":"d26b44c0a4991234ca01809fde23785ebadadd664459fdfdcc3f5d3caee4e043"} Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.790802 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.803985 4808 generic.go:334] "Generic (PLEG): container finished" podID="6bb3e28c-45c6-4992-beee-69dfd1ae7d55" containerID="fdaba41a61c2b3d0581b92ebe8d623669bd035a74fb1318c3113c88694e1bec3" exitCode=0 Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.804148 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lzcv" event={"ID":"6bb3e28c-45c6-4992-beee-69dfd1ae7d55","Type":"ContainerDied","Data":"fdaba41a61c2b3d0581b92ebe8d623669bd035a74fb1318c3113c88694e1bec3"} Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.813540 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7wlcv" event={"ID":"33777903-22e8-4520-a301-53aa0fbc9ef2","Type":"ContainerStarted","Data":"95f86bebfac4ddf5ad04357b5741a8b501fb8e04e71d42050d3a7c01b6e7696a"} Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.813675 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7wlcv" podUID="33777903-22e8-4520-a301-53aa0fbc9ef2" containerName="extract-content" containerID="cri-o://95f86bebfac4ddf5ad04357b5741a8b501fb8e04e71d42050d3a7c01b6e7696a" gracePeriod=30 Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.830557 4808 generic.go:334] "Generic (PLEG): container finished" podID="0691f390-5c98-4c03-9839-e72ba2e6c947" containerID="876190008b1baf4c11d3ccc27459e0edfd12547852e0f22cd679bc0cc83cfe74" exitCode=0 Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.830792 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79r4z" event={"ID":"0691f390-5c98-4c03-9839-e72ba2e6c947","Type":"ContainerDied","Data":"876190008b1baf4c11d3ccc27459e0edfd12547852e0f22cd679bc0cc83cfe74"} Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.859105 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" podStartSLOduration=11.859085875 podStartE2EDuration="11.859085875s" podCreationTimestamp="2026-01-21 17:59:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 18:00:04.838145146 +0000 UTC m=+239.568656071" watchObservedRunningTime="2026-01-21 18:00:04.859085875 +0000 UTC m=+239.589596760" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.876179 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdpdl" event={"ID":"f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6","Type":"ContainerStarted","Data":"44c9a9b0d875fd6d3c0feb71a350e5aa5f8a0f9337a2c3fd035ae3ac1ec301da"} Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.881047 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jdpdl" podUID="f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6" containerName="extract-content" containerID="cri-o://44c9a9b0d875fd6d3c0feb71a350e5aa5f8a0f9337a2c3fd035ae3ac1ec301da" gracePeriod=30 Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.882715 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" event={"ID":"ae9be85d-1887-4d2b-95ad-a71cdd7f8f10","Type":"ContainerStarted","Data":"00c1f4a75e8f3ab0323a0a4797b293d4652edc5c5a12acdfa7fbb9facf6d0c4e"} Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.887656 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cklmp" event={"ID":"0a6a5330-18dc-4ab0-bfb0-6b98549e28bc","Type":"ContainerStarted","Data":"656289e99d435ca7f59cdae32a1e6c3275a20c17007b1c62dc458c1141b76992"} Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.888417 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-cklmp" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.899332 4808 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cklmp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.58:8080/healthz\": dial tcp 10.217.0.58:8080: connect: connection refused" start-of-body= Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.899396 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cklmp" podUID="0a6a5330-18dc-4ab0-bfb0-6b98549e28bc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.58:8080/healthz\": dial tcp 10.217.0.58:8080: connect: connection refused" Jan 21 18:00:04 crc kubenswrapper[4808]: I0121 18:00:04.919736 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-cklmp" podStartSLOduration=3.919708765 podStartE2EDuration="3.919708765s" podCreationTimestamp="2026-01-21 18:00:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 18:00:04.917688604 +0000 UTC m=+239.648199499" watchObservedRunningTime="2026-01-21 18:00:04.919708765 +0000 UTC m=+239.650219650" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.003130 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" podStartSLOduration=5.003105769 podStartE2EDuration="5.003105769s" podCreationTimestamp="2026-01-21 18:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 18:00:04.963065809 +0000 UTC m=+239.693576694" watchObservedRunningTime="2026-01-21 18:00:05.003105769 +0000 UTC m=+239.733616654" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.018862 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lx72r"] Jan 21 18:00:05 crc kubenswrapper[4808]: W0121 18:00:05.074341 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb05e5192_b448_446a_96ba_3ec53d3644bf.slice/crio-27d08e2d789e1f390fc3e64743f37ba3962f0a028c36ff21c630239b6dcb13ef WatchSource:0}: Error finding container 27d08e2d789e1f390fc3e64743f37ba3962f0a028c36ff21c630239b6dcb13ef: Status 404 returned error can't find the container with id 27d08e2d789e1f390fc3e64743f37ba3962f0a028c36ff21c630239b6dcb13ef Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.254790 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2lzcv" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.256504 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-79r4z" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.276663 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bb3e28c-45c6-4992-beee-69dfd1ae7d55-catalog-content\") pod \"6bb3e28c-45c6-4992-beee-69dfd1ae7d55\" (UID: \"6bb3e28c-45c6-4992-beee-69dfd1ae7d55\") " Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.276745 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjb9v\" (UniqueName: \"kubernetes.io/projected/0691f390-5c98-4c03-9839-e72ba2e6c947-kube-api-access-rjb9v\") pod \"0691f390-5c98-4c03-9839-e72ba2e6c947\" (UID: \"0691f390-5c98-4c03-9839-e72ba2e6c947\") " Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.276838 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6f5w\" (UniqueName: \"kubernetes.io/projected/6bb3e28c-45c6-4992-beee-69dfd1ae7d55-kube-api-access-m6f5w\") pod \"6bb3e28c-45c6-4992-beee-69dfd1ae7d55\" (UID: \"6bb3e28c-45c6-4992-beee-69dfd1ae7d55\") " Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.276870 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bb3e28c-45c6-4992-beee-69dfd1ae7d55-utilities\") pod \"6bb3e28c-45c6-4992-beee-69dfd1ae7d55\" (UID: \"6bb3e28c-45c6-4992-beee-69dfd1ae7d55\") " Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.276904 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0691f390-5c98-4c03-9839-e72ba2e6c947-catalog-content\") pod \"0691f390-5c98-4c03-9839-e72ba2e6c947\" (UID: \"0691f390-5c98-4c03-9839-e72ba2e6c947\") " Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.276965 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0691f390-5c98-4c03-9839-e72ba2e6c947-utilities\") pod \"0691f390-5c98-4c03-9839-e72ba2e6c947\" (UID: \"0691f390-5c98-4c03-9839-e72ba2e6c947\") " Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.280027 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6bb3e28c-45c6-4992-beee-69dfd1ae7d55-utilities" (OuterVolumeSpecName: "utilities") pod "6bb3e28c-45c6-4992-beee-69dfd1ae7d55" (UID: "6bb3e28c-45c6-4992-beee-69dfd1ae7d55"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.280411 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0691f390-5c98-4c03-9839-e72ba2e6c947-utilities" (OuterVolumeSpecName: "utilities") pod "0691f390-5c98-4c03-9839-e72ba2e6c947" (UID: "0691f390-5c98-4c03-9839-e72ba2e6c947"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.287711 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bb3e28c-45c6-4992-beee-69dfd1ae7d55-kube-api-access-m6f5w" (OuterVolumeSpecName: "kube-api-access-m6f5w") pod "6bb3e28c-45c6-4992-beee-69dfd1ae7d55" (UID: "6bb3e28c-45c6-4992-beee-69dfd1ae7d55"). InnerVolumeSpecName "kube-api-access-m6f5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.287855 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6f5w\" (UniqueName: \"kubernetes.io/projected/6bb3e28c-45c6-4992-beee-69dfd1ae7d55-kube-api-access-m6f5w\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.287879 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bb3e28c-45c6-4992-beee-69dfd1ae7d55-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.287893 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0691f390-5c98-4c03-9839-e72ba2e6c947-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.299947 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0691f390-5c98-4c03-9839-e72ba2e6c947-kube-api-access-rjb9v" (OuterVolumeSpecName: "kube-api-access-rjb9v") pod "0691f390-5c98-4c03-9839-e72ba2e6c947" (UID: "0691f390-5c98-4c03-9839-e72ba2e6c947"). InnerVolumeSpecName "kube-api-access-rjb9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.312199 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0691f390-5c98-4c03-9839-e72ba2e6c947-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0691f390-5c98-4c03-9839-e72ba2e6c947" (UID: "0691f390-5c98-4c03-9839-e72ba2e6c947"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.325333 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4fsv9_3a84f07d-c20d-4a1d-a272-118d29bcdb60/extract-content/0.log" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.325743 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4fsv9" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.331082 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7wlcv_33777903-22e8-4520-a301-53aa0fbc9ef2/extract-content/0.log" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.331570 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7wlcv" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.336863 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jdpdl_f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6/extract-content/0.log" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.337432 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jdpdl" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.368680 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6bb3e28c-45c6-4992-beee-69dfd1ae7d55-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6bb3e28c-45c6-4992-beee-69dfd1ae7d55" (UID: "6bb3e28c-45c6-4992-beee-69dfd1ae7d55"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.390651 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6-catalog-content\") pod \"f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6\" (UID: \"f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6\") " Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.390748 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7zss\" (UniqueName: \"kubernetes.io/projected/33777903-22e8-4520-a301-53aa0fbc9ef2-kube-api-access-r7zss\") pod \"33777903-22e8-4520-a301-53aa0fbc9ef2\" (UID: \"33777903-22e8-4520-a301-53aa0fbc9ef2\") " Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.390789 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a84f07d-c20d-4a1d-a272-118d29bcdb60-catalog-content\") pod \"3a84f07d-c20d-4a1d-a272-118d29bcdb60\" (UID: \"3a84f07d-c20d-4a1d-a272-118d29bcdb60\") " Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.390815 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tggj\" (UniqueName: \"kubernetes.io/projected/3a84f07d-c20d-4a1d-a272-118d29bcdb60-kube-api-access-6tggj\") pod \"3a84f07d-c20d-4a1d-a272-118d29bcdb60\" (UID: \"3a84f07d-c20d-4a1d-a272-118d29bcdb60\") " Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.390845 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33777903-22e8-4520-a301-53aa0fbc9ef2-utilities\") pod \"33777903-22e8-4520-a301-53aa0fbc9ef2\" (UID: \"33777903-22e8-4520-a301-53aa0fbc9ef2\") " Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.391398 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6-utilities\") pod \"f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6\" (UID: \"f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6\") " Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.391461 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a84f07d-c20d-4a1d-a272-118d29bcdb60-utilities\") pod \"3a84f07d-c20d-4a1d-a272-118d29bcdb60\" (UID: \"3a84f07d-c20d-4a1d-a272-118d29bcdb60\") " Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.391518 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhj8j\" (UniqueName: \"kubernetes.io/projected/f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6-kube-api-access-fhj8j\") pod \"f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6\" (UID: \"f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6\") " Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.391549 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33777903-22e8-4520-a301-53aa0fbc9ef2-catalog-content\") pod \"33777903-22e8-4520-a301-53aa0fbc9ef2\" (UID: \"33777903-22e8-4520-a301-53aa0fbc9ef2\") " Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.391932 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0691f390-5c98-4c03-9839-e72ba2e6c947-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.393937 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6-utilities" (OuterVolumeSpecName: "utilities") pod "f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6" (UID: "f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.394598 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bb3e28c-45c6-4992-beee-69dfd1ae7d55-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.394664 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjb9v\" (UniqueName: \"kubernetes.io/projected/0691f390-5c98-4c03-9839-e72ba2e6c947-kube-api-access-rjb9v\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.401157 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33777903-22e8-4520-a301-53aa0fbc9ef2-utilities" (OuterVolumeSpecName: "utilities") pod "33777903-22e8-4520-a301-53aa0fbc9ef2" (UID: "33777903-22e8-4520-a301-53aa0fbc9ef2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.401362 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a84f07d-c20d-4a1d-a272-118d29bcdb60-utilities" (OuterVolumeSpecName: "utilities") pod "3a84f07d-c20d-4a1d-a272-118d29bcdb60" (UID: "3a84f07d-c20d-4a1d-a272-118d29bcdb60"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.405112 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6-kube-api-access-fhj8j" (OuterVolumeSpecName: "kube-api-access-fhj8j") pod "f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6" (UID: "f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6"). InnerVolumeSpecName "kube-api-access-fhj8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.408846 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33777903-22e8-4520-a301-53aa0fbc9ef2-kube-api-access-r7zss" (OuterVolumeSpecName: "kube-api-access-r7zss") pod "33777903-22e8-4520-a301-53aa0fbc9ef2" (UID: "33777903-22e8-4520-a301-53aa0fbc9ef2"). InnerVolumeSpecName "kube-api-access-r7zss". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.423296 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a84f07d-c20d-4a1d-a272-118d29bcdb60-kube-api-access-6tggj" (OuterVolumeSpecName: "kube-api-access-6tggj") pod "3a84f07d-c20d-4a1d-a272-118d29bcdb60" (UID: "3a84f07d-c20d-4a1d-a272-118d29bcdb60"). InnerVolumeSpecName "kube-api-access-6tggj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.440112 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6" (UID: "f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.440579 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a84f07d-c20d-4a1d-a272-118d29bcdb60-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a84f07d-c20d-4a1d-a272-118d29bcdb60" (UID: "3a84f07d-c20d-4a1d-a272-118d29bcdb60"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.457249 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33777903-22e8-4520-a301-53aa0fbc9ef2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "33777903-22e8-4520-a301-53aa0fbc9ef2" (UID: "33777903-22e8-4520-a301-53aa0fbc9ef2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.495740 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7zss\" (UniqueName: \"kubernetes.io/projected/33777903-22e8-4520-a301-53aa0fbc9ef2-kube-api-access-r7zss\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.495774 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a84f07d-c20d-4a1d-a272-118d29bcdb60-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.495784 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tggj\" (UniqueName: \"kubernetes.io/projected/3a84f07d-c20d-4a1d-a272-118d29bcdb60-kube-api-access-6tggj\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.495824 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33777903-22e8-4520-a301-53aa0fbc9ef2-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.495838 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.495848 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a84f07d-c20d-4a1d-a272-118d29bcdb60-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.495858 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhj8j\" (UniqueName: \"kubernetes.io/projected/f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6-kube-api-access-fhj8j\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.495867 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33777903-22e8-4520-a301-53aa0fbc9ef2-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.495875 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.528779 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09939aec-8ef7-4b4d-a934-9f00e19bc7ed" path="/var/lib/kubelet/pods/09939aec-8ef7-4b4d-a934-9f00e19bc7ed/volumes" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.530340 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="595fab48-bf4b-443f-9367-d5718a4672e3" path="/var/lib/kubelet/pods/595fab48-bf4b-443f-9367-d5718a4672e3/volumes" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.532838 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5" path="/var/lib/kubelet/pods/5a8b67b7-9e47-4bad-881e-7b8c2ab53ca5/volumes" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.534542 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6221711c-1051-44e9-8fad-029c73d0ae71" path="/var/lib/kubelet/pods/6221711c-1051-44e9-8fad-029c73d0ae71/volumes" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.895336 4808 generic.go:334] "Generic (PLEG): container finished" podID="33777903-22e8-4520-a301-53aa0fbc9ef2" containerID="95f86bebfac4ddf5ad04357b5741a8b501fb8e04e71d42050d3a7c01b6e7696a" exitCode=2 Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.895420 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7wlcv" event={"ID":"33777903-22e8-4520-a301-53aa0fbc9ef2","Type":"ContainerDied","Data":"95f86bebfac4ddf5ad04357b5741a8b501fb8e04e71d42050d3a7c01b6e7696a"} Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.895434 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7wlcv" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.895456 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7wlcv" event={"ID":"33777903-22e8-4520-a301-53aa0fbc9ef2","Type":"ContainerDied","Data":"7e69469f544bead1148dd283bddb9d4769bda103dee8d1f8fb5a0eddfa8b979d"} Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.895478 4808 scope.go:117] "RemoveContainer" containerID="95f86bebfac4ddf5ad04357b5741a8b501fb8e04e71d42050d3a7c01b6e7696a" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.897009 4808 generic.go:334] "Generic (PLEG): container finished" podID="b05e5192-b448-446a-96ba-3ec53d3644bf" containerID="568239dfedb5eb527404b60490c0913fe635260d28ee528f39bb6e9f8f6f4e05" exitCode=0 Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.897085 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lx72r" event={"ID":"b05e5192-b448-446a-96ba-3ec53d3644bf","Type":"ContainerDied","Data":"568239dfedb5eb527404b60490c0913fe635260d28ee528f39bb6e9f8f6f4e05"} Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.897130 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lx72r" event={"ID":"b05e5192-b448-446a-96ba-3ec53d3644bf","Type":"ContainerStarted","Data":"27d08e2d789e1f390fc3e64743f37ba3962f0a028c36ff21c630239b6dcb13ef"} Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.903439 4808 generic.go:334] "Generic (PLEG): container finished" podID="3a84f07d-c20d-4a1d-a272-118d29bcdb60" containerID="917d69008a1ee33f31200c809e8c9a31224d1fdb26bd388c4c074af8d0d90ab7" exitCode=2 Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.903523 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4fsv9" event={"ID":"3a84f07d-c20d-4a1d-a272-118d29bcdb60","Type":"ContainerDied","Data":"917d69008a1ee33f31200c809e8c9a31224d1fdb26bd388c4c074af8d0d90ab7"} Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.903557 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4fsv9" event={"ID":"3a84f07d-c20d-4a1d-a272-118d29bcdb60","Type":"ContainerDied","Data":"4661d49dbe78a3ff03c5f1fcc3258eb33d61c9501d34d336285b2760c00bfc35"} Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.903568 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4fsv9" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.908948 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79r4z" event={"ID":"0691f390-5c98-4c03-9839-e72ba2e6c947","Type":"ContainerDied","Data":"5268301c4d84f88fc2fca5fd9a5111338717f22a6b5908ca66abc98c3be54ec7"} Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.909221 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-79r4z" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.912345 4808 generic.go:334] "Generic (PLEG): container finished" podID="f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6" containerID="44c9a9b0d875fd6d3c0feb71a350e5aa5f8a0f9337a2c3fd035ae3ac1ec301da" exitCode=2 Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.912419 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdpdl" event={"ID":"f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6","Type":"ContainerDied","Data":"44c9a9b0d875fd6d3c0feb71a350e5aa5f8a0f9337a2c3fd035ae3ac1ec301da"} Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.912448 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdpdl" event={"ID":"f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6","Type":"ContainerDied","Data":"d50f734f9698fdeac0f9c10ff2d9923763b952bf7feaf382c3f84c71a3c4ee68"} Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.912535 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jdpdl" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.915171 4808 generic.go:334] "Generic (PLEG): container finished" podID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" containerID="00c1f4a75e8f3ab0323a0a4797b293d4652edc5c5a12acdfa7fbb9facf6d0c4e" exitCode=0 Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.915271 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" event={"ID":"ae9be85d-1887-4d2b-95ad-a71cdd7f8f10","Type":"ContainerDied","Data":"00c1f4a75e8f3ab0323a0a4797b293d4652edc5c5a12acdfa7fbb9facf6d0c4e"} Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.917732 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2lzcv" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.918287 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lzcv" event={"ID":"6bb3e28c-45c6-4992-beee-69dfd1ae7d55","Type":"ContainerDied","Data":"ed0f63879b8d9fda79a6c605be730962090c2c91ed5a5d5813f707491810975a"} Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.929451 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-cklmp" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.957970 4808 scope.go:117] "RemoveContainer" containerID="29df6527d44100973e0c8c31d811b6c7902efd577921ffc46fb4cb8f2812694d" Jan 21 18:00:05 crc kubenswrapper[4808]: I0121 18:00:05.992545 4808 scope.go:117] "RemoveContainer" containerID="95f86bebfac4ddf5ad04357b5741a8b501fb8e04e71d42050d3a7c01b6e7696a" Jan 21 18:00:06 crc kubenswrapper[4808]: E0121 18:00:06.007002 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95f86bebfac4ddf5ad04357b5741a8b501fb8e04e71d42050d3a7c01b6e7696a\": container with ID starting with 95f86bebfac4ddf5ad04357b5741a8b501fb8e04e71d42050d3a7c01b6e7696a not found: ID does not exist" containerID="95f86bebfac4ddf5ad04357b5741a8b501fb8e04e71d42050d3a7c01b6e7696a" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.007050 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95f86bebfac4ddf5ad04357b5741a8b501fb8e04e71d42050d3a7c01b6e7696a"} err="failed to get container status \"95f86bebfac4ddf5ad04357b5741a8b501fb8e04e71d42050d3a7c01b6e7696a\": rpc error: code = NotFound desc = could not find container \"95f86bebfac4ddf5ad04357b5741a8b501fb8e04e71d42050d3a7c01b6e7696a\": container with ID starting with 95f86bebfac4ddf5ad04357b5741a8b501fb8e04e71d42050d3a7c01b6e7696a not found: ID does not exist" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.007106 4808 scope.go:117] "RemoveContainer" containerID="29df6527d44100973e0c8c31d811b6c7902efd577921ffc46fb4cb8f2812694d" Jan 21 18:00:06 crc kubenswrapper[4808]: E0121 18:00:06.007706 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29df6527d44100973e0c8c31d811b6c7902efd577921ffc46fb4cb8f2812694d\": container with ID starting with 29df6527d44100973e0c8c31d811b6c7902efd577921ffc46fb4cb8f2812694d not found: ID does not exist" containerID="29df6527d44100973e0c8c31d811b6c7902efd577921ffc46fb4cb8f2812694d" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.007747 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29df6527d44100973e0c8c31d811b6c7902efd577921ffc46fb4cb8f2812694d"} err="failed to get container status \"29df6527d44100973e0c8c31d811b6c7902efd577921ffc46fb4cb8f2812694d\": rpc error: code = NotFound desc = could not find container \"29df6527d44100973e0c8c31d811b6c7902efd577921ffc46fb4cb8f2812694d\": container with ID starting with 29df6527d44100973e0c8c31d811b6c7902efd577921ffc46fb4cb8f2812694d not found: ID does not exist" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.007771 4808 scope.go:117] "RemoveContainer" containerID="917d69008a1ee33f31200c809e8c9a31224d1fdb26bd388c4c074af8d0d90ab7" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.030149 4808 scope.go:117] "RemoveContainer" containerID="8032d1838a45c097010150327581f76f18bcef8e58d62811ebccf52728ff3f8f" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.035809 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7wlcv"] Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.037974 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7wlcv"] Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.052518 4808 scope.go:117] "RemoveContainer" containerID="917d69008a1ee33f31200c809e8c9a31224d1fdb26bd388c4c074af8d0d90ab7" Jan 21 18:00:06 crc kubenswrapper[4808]: E0121 18:00:06.053774 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"917d69008a1ee33f31200c809e8c9a31224d1fdb26bd388c4c074af8d0d90ab7\": container with ID starting with 917d69008a1ee33f31200c809e8c9a31224d1fdb26bd388c4c074af8d0d90ab7 not found: ID does not exist" containerID="917d69008a1ee33f31200c809e8c9a31224d1fdb26bd388c4c074af8d0d90ab7" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.053805 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"917d69008a1ee33f31200c809e8c9a31224d1fdb26bd388c4c074af8d0d90ab7"} err="failed to get container status \"917d69008a1ee33f31200c809e8c9a31224d1fdb26bd388c4c074af8d0d90ab7\": rpc error: code = NotFound desc = could not find container \"917d69008a1ee33f31200c809e8c9a31224d1fdb26bd388c4c074af8d0d90ab7\": container with ID starting with 917d69008a1ee33f31200c809e8c9a31224d1fdb26bd388c4c074af8d0d90ab7 not found: ID does not exist" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.053829 4808 scope.go:117] "RemoveContainer" containerID="8032d1838a45c097010150327581f76f18bcef8e58d62811ebccf52728ff3f8f" Jan 21 18:00:06 crc kubenswrapper[4808]: E0121 18:00:06.054227 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8032d1838a45c097010150327581f76f18bcef8e58d62811ebccf52728ff3f8f\": container with ID starting with 8032d1838a45c097010150327581f76f18bcef8e58d62811ebccf52728ff3f8f not found: ID does not exist" containerID="8032d1838a45c097010150327581f76f18bcef8e58d62811ebccf52728ff3f8f" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.054294 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8032d1838a45c097010150327581f76f18bcef8e58d62811ebccf52728ff3f8f"} err="failed to get container status \"8032d1838a45c097010150327581f76f18bcef8e58d62811ebccf52728ff3f8f\": rpc error: code = NotFound desc = could not find container \"8032d1838a45c097010150327581f76f18bcef8e58d62811ebccf52728ff3f8f\": container with ID starting with 8032d1838a45c097010150327581f76f18bcef8e58d62811ebccf52728ff3f8f not found: ID does not exist" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.054332 4808 scope.go:117] "RemoveContainer" containerID="876190008b1baf4c11d3ccc27459e0edfd12547852e0f22cd679bc0cc83cfe74" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.066802 4808 scope.go:117] "RemoveContainer" containerID="80ffabddaa3f7488c74769be311fc5a95bc4479e0947d9dbfb757d1c0589cbd2" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.094509 4808 scope.go:117] "RemoveContainer" containerID="44c9a9b0d875fd6d3c0feb71a350e5aa5f8a0f9337a2c3fd035ae3ac1ec301da" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.096183 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jdpdl"] Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.100160 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jdpdl"] Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.112765 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4fsv9"] Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.133152 4808 scope.go:117] "RemoveContainer" containerID="61bc74b4c9a7b0890598e054c11076b36a23c5835b24148a62486c337c8d9360" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.136844 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4fsv9"] Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.149778 4808 scope.go:117] "RemoveContainer" containerID="44c9a9b0d875fd6d3c0feb71a350e5aa5f8a0f9337a2c3fd035ae3ac1ec301da" Jan 21 18:00:06 crc kubenswrapper[4808]: E0121 18:00:06.150205 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44c9a9b0d875fd6d3c0feb71a350e5aa5f8a0f9337a2c3fd035ae3ac1ec301da\": container with ID starting with 44c9a9b0d875fd6d3c0feb71a350e5aa5f8a0f9337a2c3fd035ae3ac1ec301da not found: ID does not exist" containerID="44c9a9b0d875fd6d3c0feb71a350e5aa5f8a0f9337a2c3fd035ae3ac1ec301da" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.150276 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44c9a9b0d875fd6d3c0feb71a350e5aa5f8a0f9337a2c3fd035ae3ac1ec301da"} err="failed to get container status \"44c9a9b0d875fd6d3c0feb71a350e5aa5f8a0f9337a2c3fd035ae3ac1ec301da\": rpc error: code = NotFound desc = could not find container \"44c9a9b0d875fd6d3c0feb71a350e5aa5f8a0f9337a2c3fd035ae3ac1ec301da\": container with ID starting with 44c9a9b0d875fd6d3c0feb71a350e5aa5f8a0f9337a2c3fd035ae3ac1ec301da not found: ID does not exist" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.150307 4808 scope.go:117] "RemoveContainer" containerID="61bc74b4c9a7b0890598e054c11076b36a23c5835b24148a62486c337c8d9360" Jan 21 18:00:06 crc kubenswrapper[4808]: E0121 18:00:06.150771 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61bc74b4c9a7b0890598e054c11076b36a23c5835b24148a62486c337c8d9360\": container with ID starting with 61bc74b4c9a7b0890598e054c11076b36a23c5835b24148a62486c337c8d9360 not found: ID does not exist" containerID="61bc74b4c9a7b0890598e054c11076b36a23c5835b24148a62486c337c8d9360" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.150791 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61bc74b4c9a7b0890598e054c11076b36a23c5835b24148a62486c337c8d9360"} err="failed to get container status \"61bc74b4c9a7b0890598e054c11076b36a23c5835b24148a62486c337c8d9360\": rpc error: code = NotFound desc = could not find container \"61bc74b4c9a7b0890598e054c11076b36a23c5835b24148a62486c337c8d9360\": container with ID starting with 61bc74b4c9a7b0890598e054c11076b36a23c5835b24148a62486c337c8d9360 not found: ID does not exist" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.150804 4808 scope.go:117] "RemoveContainer" containerID="fdaba41a61c2b3d0581b92ebe8d623669bd035a74fb1318c3113c88694e1bec3" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.168102 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-79r4z"] Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.174524 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-79r4z"] Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.179390 4808 scope.go:117] "RemoveContainer" containerID="5d7a6b63cf66a4c74602ce95aac4a7cdbebba8d0d5c653c14a45cda28984c7e4" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.199580 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2lzcv"] Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.202475 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2lzcv"] Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.761713 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mzpzk"] Jan 21 18:00:06 crc kubenswrapper[4808]: E0121 18:00:06.762449 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a84f07d-c20d-4a1d-a272-118d29bcdb60" containerName="extract-content" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.762470 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a84f07d-c20d-4a1d-a272-118d29bcdb60" containerName="extract-content" Jan 21 18:00:06 crc kubenswrapper[4808]: E0121 18:00:06.762491 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33777903-22e8-4520-a301-53aa0fbc9ef2" containerName="extract-content" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.762504 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="33777903-22e8-4520-a301-53aa0fbc9ef2" containerName="extract-content" Jan 21 18:00:06 crc kubenswrapper[4808]: E0121 18:00:06.762523 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0691f390-5c98-4c03-9839-e72ba2e6c947" containerName="extract-utilities" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.762537 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="0691f390-5c98-4c03-9839-e72ba2e6c947" containerName="extract-utilities" Jan 21 18:00:06 crc kubenswrapper[4808]: E0121 18:00:06.762559 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a84f07d-c20d-4a1d-a272-118d29bcdb60" containerName="extract-utilities" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.762571 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a84f07d-c20d-4a1d-a272-118d29bcdb60" containerName="extract-utilities" Jan 21 18:00:06 crc kubenswrapper[4808]: E0121 18:00:06.762587 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33777903-22e8-4520-a301-53aa0fbc9ef2" containerName="extract-utilities" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.762600 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="33777903-22e8-4520-a301-53aa0fbc9ef2" containerName="extract-utilities" Jan 21 18:00:06 crc kubenswrapper[4808]: E0121 18:00:06.762619 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bb3e28c-45c6-4992-beee-69dfd1ae7d55" containerName="extract-utilities" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.762632 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bb3e28c-45c6-4992-beee-69dfd1ae7d55" containerName="extract-utilities" Jan 21 18:00:06 crc kubenswrapper[4808]: E0121 18:00:06.762653 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bb3e28c-45c6-4992-beee-69dfd1ae7d55" containerName="extract-content" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.762665 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bb3e28c-45c6-4992-beee-69dfd1ae7d55" containerName="extract-content" Jan 21 18:00:06 crc kubenswrapper[4808]: E0121 18:00:06.762682 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0691f390-5c98-4c03-9839-e72ba2e6c947" containerName="extract-content" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.762695 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="0691f390-5c98-4c03-9839-e72ba2e6c947" containerName="extract-content" Jan 21 18:00:06 crc kubenswrapper[4808]: E0121 18:00:06.762713 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6" containerName="extract-content" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.762724 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6" containerName="extract-content" Jan 21 18:00:06 crc kubenswrapper[4808]: E0121 18:00:06.762744 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6" containerName="extract-utilities" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.762756 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6" containerName="extract-utilities" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.762919 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a84f07d-c20d-4a1d-a272-118d29bcdb60" containerName="extract-content" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.762941 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6" containerName="extract-content" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.762963 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="33777903-22e8-4520-a301-53aa0fbc9ef2" containerName="extract-content" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.762979 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="0691f390-5c98-4c03-9839-e72ba2e6c947" containerName="extract-content" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.762995 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bb3e28c-45c6-4992-beee-69dfd1ae7d55" containerName="extract-content" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.764287 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mzpzk" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.767838 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.778282 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mzpzk"] Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.831917 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rhxt\" (UniqueName: \"kubernetes.io/projected/688d7727-51ad-456f-a651-cea655ee7507-kube-api-access-4rhxt\") pod \"community-operators-mzpzk\" (UID: \"688d7727-51ad-456f-a651-cea655ee7507\") " pod="openshift-marketplace/community-operators-mzpzk" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.831966 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/688d7727-51ad-456f-a651-cea655ee7507-utilities\") pod \"community-operators-mzpzk\" (UID: \"688d7727-51ad-456f-a651-cea655ee7507\") " pod="openshift-marketplace/community-operators-mzpzk" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.831998 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/688d7727-51ad-456f-a651-cea655ee7507-catalog-content\") pod \"community-operators-mzpzk\" (UID: \"688d7727-51ad-456f-a651-cea655ee7507\") " pod="openshift-marketplace/community-operators-mzpzk" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.933337 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rhxt\" (UniqueName: \"kubernetes.io/projected/688d7727-51ad-456f-a651-cea655ee7507-kube-api-access-4rhxt\") pod \"community-operators-mzpzk\" (UID: \"688d7727-51ad-456f-a651-cea655ee7507\") " pod="openshift-marketplace/community-operators-mzpzk" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.933849 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/688d7727-51ad-456f-a651-cea655ee7507-utilities\") pod \"community-operators-mzpzk\" (UID: \"688d7727-51ad-456f-a651-cea655ee7507\") " pod="openshift-marketplace/community-operators-mzpzk" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.933957 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/688d7727-51ad-456f-a651-cea655ee7507-catalog-content\") pod \"community-operators-mzpzk\" (UID: \"688d7727-51ad-456f-a651-cea655ee7507\") " pod="openshift-marketplace/community-operators-mzpzk" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.934512 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/688d7727-51ad-456f-a651-cea655ee7507-utilities\") pod \"community-operators-mzpzk\" (UID: \"688d7727-51ad-456f-a651-cea655ee7507\") " pod="openshift-marketplace/community-operators-mzpzk" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.934568 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/688d7727-51ad-456f-a651-cea655ee7507-catalog-content\") pod \"community-operators-mzpzk\" (UID: \"688d7727-51ad-456f-a651-cea655ee7507\") " pod="openshift-marketplace/community-operators-mzpzk" Jan 21 18:00:06 crc kubenswrapper[4808]: I0121 18:00:06.955988 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rhxt\" (UniqueName: \"kubernetes.io/projected/688d7727-51ad-456f-a651-cea655ee7507-kube-api-access-4rhxt\") pod \"community-operators-mzpzk\" (UID: \"688d7727-51ad-456f-a651-cea655ee7507\") " pod="openshift-marketplace/community-operators-mzpzk" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.080226 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mzpzk" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.161760 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.179791 4808 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 21 18:00:07 crc kubenswrapper[4808]: E0121 18:00:07.180185 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" containerName="collect-profiles" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.180201 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" containerName="collect-profiles" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.180344 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" containerName="collect-profiles" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.180812 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.182001 4808 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.182405 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159" gracePeriod=15 Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.182560 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51" gracePeriod=15 Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.182606 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7" gracePeriod=15 Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.182643 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6" gracePeriod=15 Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.182673 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1" gracePeriod=15 Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.187394 4808 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 21 18:00:07 crc kubenswrapper[4808]: E0121 18:00:07.187587 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.187598 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 21 18:00:07 crc kubenswrapper[4808]: E0121 18:00:07.187607 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.187612 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 21 18:00:07 crc kubenswrapper[4808]: E0121 18:00:07.187626 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.187633 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 21 18:00:07 crc kubenswrapper[4808]: E0121 18:00:07.187640 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.187646 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 21 18:00:07 crc kubenswrapper[4808]: E0121 18:00:07.187653 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.187659 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 21 18:00:07 crc kubenswrapper[4808]: E0121 18:00:07.187667 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.187673 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 21 18:00:07 crc kubenswrapper[4808]: E0121 18:00:07.187806 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.187817 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.187930 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.187943 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.187953 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.187967 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.187976 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.188158 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.225569 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.237688 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ae9be85d-1887-4d2b-95ad-a71cdd7f8f10-secret-volume\") pod \"ae9be85d-1887-4d2b-95ad-a71cdd7f8f10\" (UID: \"ae9be85d-1887-4d2b-95ad-a71cdd7f8f10\") " Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.237767 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae9be85d-1887-4d2b-95ad-a71cdd7f8f10-config-volume\") pod \"ae9be85d-1887-4d2b-95ad-a71cdd7f8f10\" (UID: \"ae9be85d-1887-4d2b-95ad-a71cdd7f8f10\") " Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.237805 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29zqb\" (UniqueName: \"kubernetes.io/projected/ae9be85d-1887-4d2b-95ad-a71cdd7f8f10-kube-api-access-29zqb\") pod \"ae9be85d-1887-4d2b-95ad-a71cdd7f8f10\" (UID: \"ae9be85d-1887-4d2b-95ad-a71cdd7f8f10\") " Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.238078 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.238143 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.238181 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.238214 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.238303 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.238371 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.238434 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.238458 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.238714 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae9be85d-1887-4d2b-95ad-a71cdd7f8f10-config-volume" (OuterVolumeSpecName: "config-volume") pod "ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" (UID: "ae9be85d-1887-4d2b-95ad-a71cdd7f8f10"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.243029 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae9be85d-1887-4d2b-95ad-a71cdd7f8f10-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" (UID: "ae9be85d-1887-4d2b-95ad-a71cdd7f8f10"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.243336 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae9be85d-1887-4d2b-95ad-a71cdd7f8f10-kube-api-access-29zqb" (OuterVolumeSpecName: "kube-api-access-29zqb") pod "ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" (UID: "ae9be85d-1887-4d2b-95ad-a71cdd7f8f10"). InnerVolumeSpecName "kube-api-access-29zqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.274870 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mzpzk"] Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.340163 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.340230 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.340278 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.340314 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.340331 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.340355 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.340383 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.340419 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.340454 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.340449 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.340376 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.340503 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.340677 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.340797 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.340921 4808 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae9be85d-1887-4d2b-95ad-a71cdd7f8f10-config-volume\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.340956 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29zqb\" (UniqueName: \"kubernetes.io/projected/ae9be85d-1887-4d2b-95ad-a71cdd7f8f10-kube-api-access-29zqb\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.340974 4808 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ae9be85d-1887-4d2b-95ad-a71cdd7f8f10-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.341014 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.341046 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.522139 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.534882 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0691f390-5c98-4c03-9839-e72ba2e6c947" path="/var/lib/kubelet/pods/0691f390-5c98-4c03-9839-e72ba2e6c947/volumes" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.536809 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33777903-22e8-4520-a301-53aa0fbc9ef2" path="/var/lib/kubelet/pods/33777903-22e8-4520-a301-53aa0fbc9ef2/volumes" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.538284 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a84f07d-c20d-4a1d-a272-118d29bcdb60" path="/var/lib/kubelet/pods/3a84f07d-c20d-4a1d-a272-118d29bcdb60/volumes" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.540519 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bb3e28c-45c6-4992-beee-69dfd1ae7d55" path="/var/lib/kubelet/pods/6bb3e28c-45c6-4992-beee-69dfd1ae7d55/volumes" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.541646 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6" path="/var/lib/kubelet/pods/f53dff6f-bb26-4c73-b0be-5a7ad54f6cb6/volumes" Jan 21 18:00:07 crc kubenswrapper[4808]: W0121 18:00:07.554696 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-98308b3a9c7684cf618bfd78aaf63bf60d7a8b95e929b5fe615ecaa7ea13abb7 WatchSource:0}: Error finding container 98308b3a9c7684cf618bfd78aaf63bf60d7a8b95e929b5fe615ecaa7ea13abb7: Status 404 returned error can't find the container with id 98308b3a9c7684cf618bfd78aaf63bf60d7a8b95e929b5fe615ecaa7ea13abb7 Jan 21 18:00:07 crc kubenswrapper[4808]: E0121 18:00:07.559149 4808 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.102:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188cd0de77a21d78 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-21 18:00:07.558192504 +0000 UTC m=+242.288703399,LastTimestamp:2026-01-21 18:00:07.558192504 +0000 UTC m=+242.288703399,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.935670 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mzpzk" event={"ID":"688d7727-51ad-456f-a651-cea655ee7507","Type":"ContainerStarted","Data":"e7bd18e74ade605cbb1c083d069eaaae670f2ad77dea121a22cada2aecbd20e8"} Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.935728 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mzpzk" event={"ID":"688d7727-51ad-456f-a651-cea655ee7507","Type":"ContainerStarted","Data":"41e12d50840baedb395f50726c02756aec31eee001f265b7bbb8be50bef844c8"} Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.936370 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.936610 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.941535 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.943871 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.945648 4808 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51" exitCode=0 Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.945678 4808 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7" exitCode=0 Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.945688 4808 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6" exitCode=0 Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.945701 4808 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1" exitCode=2 Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.945794 4808 scope.go:117] "RemoveContainer" containerID="15679e3847754ba7656138533cb553437ee54b3d95479b834746f470271cafe7" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.949614 4808 generic.go:334] "Generic (PLEG): container finished" podID="2f152312-7210-4e7a-b9c1-6a2942f488aa" containerID="474fc1c7bd09c01331f20a863ed7d0de5601ad2931cd0abd3dffba99dc627033" exitCode=0 Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.949686 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2f152312-7210-4e7a-b9c1-6a2942f488aa","Type":"ContainerDied","Data":"474fc1c7bd09c01331f20a863ed7d0de5601ad2931cd0abd3dffba99dc627033"} Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.950444 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.950816 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.951086 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.951487 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" event={"ID":"ae9be85d-1887-4d2b-95ad-a71cdd7f8f10","Type":"ContainerDied","Data":"72eb31190dd59a8354ced5449537c7bd199bc87e0a99ebb8bad4d14ef977414b"} Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.951515 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72eb31190dd59a8354ced5449537c7bd199bc87e0a99ebb8bad4d14ef977414b" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.951551 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.952831 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b34dffa09c0a13c935c5d49dd0f7ef7c07c1b0d9e3a687e46ca8535fb551870a"} Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.952879 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"98308b3a9c7684cf618bfd78aaf63bf60d7a8b95e929b5fe615ecaa7ea13abb7"} Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.953233 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.954453 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:07 crc kubenswrapper[4808]: I0121 18:00:07.954788 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:08 crc kubenswrapper[4808]: I0121 18:00:08.041545 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:08 crc kubenswrapper[4808]: I0121 18:00:08.041935 4808 status_manager.go:851] "Failed to get status for pod" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/collect-profiles-29483640-bx86q\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:08 crc kubenswrapper[4808]: I0121 18:00:08.042104 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:08 crc kubenswrapper[4808]: I0121 18:00:08.042295 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:08 crc kubenswrapper[4808]: I0121 18:00:08.962172 4808 generic.go:334] "Generic (PLEG): container finished" podID="688d7727-51ad-456f-a651-cea655ee7507" containerID="e7bd18e74ade605cbb1c083d069eaaae670f2ad77dea121a22cada2aecbd20e8" exitCode=0 Jan 21 18:00:08 crc kubenswrapper[4808]: I0121 18:00:08.962234 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mzpzk" event={"ID":"688d7727-51ad-456f-a651-cea655ee7507","Type":"ContainerDied","Data":"e7bd18e74ade605cbb1c083d069eaaae670f2ad77dea121a22cada2aecbd20e8"} Jan 21 18:00:08 crc kubenswrapper[4808]: I0121 18:00:08.963006 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:08 crc kubenswrapper[4808]: I0121 18:00:08.963841 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:08 crc kubenswrapper[4808]: I0121 18:00:08.964156 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:08 crc kubenswrapper[4808]: I0121 18:00:08.964551 4808 status_manager.go:851] "Failed to get status for pod" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/collect-profiles-29483640-bx86q\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:08 crc kubenswrapper[4808]: I0121 18:00:08.967223 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 21 18:00:08 crc kubenswrapper[4808]: I0121 18:00:08.970855 4808 generic.go:334] "Generic (PLEG): container finished" podID="b05e5192-b448-446a-96ba-3ec53d3644bf" containerID="00f32933828b596a7577e3722d32d68205a1c6b654d7a120a69a25051225ebc1" exitCode=0 Jan 21 18:00:08 crc kubenswrapper[4808]: I0121 18:00:08.970931 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lx72r" event={"ID":"b05e5192-b448-446a-96ba-3ec53d3644bf","Type":"ContainerDied","Data":"00f32933828b596a7577e3722d32d68205a1c6b654d7a120a69a25051225ebc1"} Jan 21 18:00:08 crc kubenswrapper[4808]: I0121 18:00:08.971465 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:08 crc kubenswrapper[4808]: I0121 18:00:08.972436 4808 status_manager.go:851] "Failed to get status for pod" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/collect-profiles-29483640-bx86q\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:08 crc kubenswrapper[4808]: I0121 18:00:08.972645 4808 status_manager.go:851] "Failed to get status for pod" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" pod="openshift-marketplace/redhat-marketplace-lx72r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-lx72r\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:08 crc kubenswrapper[4808]: I0121 18:00:08.972791 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:08 crc kubenswrapper[4808]: I0121 18:00:08.972987 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.236438 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.237215 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.237668 4808 status_manager.go:851] "Failed to get status for pod" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/collect-profiles-29483640-bx86q\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.237936 4808 status_manager.go:851] "Failed to get status for pod" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" pod="openshift-marketplace/redhat-marketplace-lx72r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-lx72r\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.238177 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.238471 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.364833 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2f152312-7210-4e7a-b9c1-6a2942f488aa-var-lock\") pod \"2f152312-7210-4e7a-b9c1-6a2942f488aa\" (UID: \"2f152312-7210-4e7a-b9c1-6a2942f488aa\") " Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.364902 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f152312-7210-4e7a-b9c1-6a2942f488aa-kube-api-access\") pod \"2f152312-7210-4e7a-b9c1-6a2942f488aa\" (UID: \"2f152312-7210-4e7a-b9c1-6a2942f488aa\") " Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.364950 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f152312-7210-4e7a-b9c1-6a2942f488aa-kubelet-dir\") pod \"2f152312-7210-4e7a-b9c1-6a2942f488aa\" (UID: \"2f152312-7210-4e7a-b9c1-6a2942f488aa\") " Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.365282 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f152312-7210-4e7a-b9c1-6a2942f488aa-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2f152312-7210-4e7a-b9c1-6a2942f488aa" (UID: "2f152312-7210-4e7a-b9c1-6a2942f488aa"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.365314 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f152312-7210-4e7a-b9c1-6a2942f488aa-var-lock" (OuterVolumeSpecName: "var-lock") pod "2f152312-7210-4e7a-b9c1-6a2942f488aa" (UID: "2f152312-7210-4e7a-b9c1-6a2942f488aa"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.372352 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f152312-7210-4e7a-b9c1-6a2942f488aa-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2f152312-7210-4e7a-b9c1-6a2942f488aa" (UID: "2f152312-7210-4e7a-b9c1-6a2942f488aa"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.466490 4808 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2f152312-7210-4e7a-b9c1-6a2942f488aa-var-lock\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.466538 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f152312-7210-4e7a-b9c1-6a2942f488aa-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.466550 4808 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f152312-7210-4e7a-b9c1-6a2942f488aa-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.978979 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2f152312-7210-4e7a-b9c1-6a2942f488aa","Type":"ContainerDied","Data":"39cd1f403cdd0ab5909b5d3b32782969ee12f74e7ffbb3231c202f91b04dadae"} Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.979386 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39cd1f403cdd0ab5909b5d3b32782969ee12f74e7ffbb3231c202f91b04dadae" Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.979461 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.983874 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.984809 4808 status_manager.go:851] "Failed to get status for pod" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/collect-profiles-29483640-bx86q\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.985712 4808 status_manager.go:851] "Failed to get status for pod" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" pod="openshift-marketplace/redhat-marketplace-lx72r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-lx72r\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.986091 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:09 crc kubenswrapper[4808]: I0121 18:00:09.986457 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:10 crc kubenswrapper[4808]: I0121 18:00:10.991252 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 21 18:00:10 crc kubenswrapper[4808]: I0121 18:00:10.992819 4808 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159" exitCode=0 Jan 21 18:00:11 crc kubenswrapper[4808]: E0121 18:00:11.374546 4808 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:11 crc kubenswrapper[4808]: E0121 18:00:11.374925 4808 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:11 crc kubenswrapper[4808]: E0121 18:00:11.375286 4808 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:11 crc kubenswrapper[4808]: E0121 18:00:11.375609 4808 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:11 crc kubenswrapper[4808]: E0121 18:00:11.375875 4808 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:11 crc kubenswrapper[4808]: I0121 18:00:11.375906 4808 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Jan 21 18:00:11 crc kubenswrapper[4808]: E0121 18:00:11.376183 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.102:6443: connect: connection refused" interval="200ms" Jan 21 18:00:11 crc kubenswrapper[4808]: I0121 18:00:11.378031 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 21 18:00:11 crc kubenswrapper[4808]: I0121 18:00:11.378983 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:11 crc kubenswrapper[4808]: I0121 18:00:11.379703 4808 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:11 crc kubenswrapper[4808]: I0121 18:00:11.380042 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:11 crc kubenswrapper[4808]: I0121 18:00:11.380446 4808 status_manager.go:851] "Failed to get status for pod" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/collect-profiles-29483640-bx86q\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:11 crc kubenswrapper[4808]: I0121 18:00:11.380725 4808 status_manager.go:851] "Failed to get status for pod" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" pod="openshift-marketplace/redhat-marketplace-lx72r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-lx72r\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:11 crc kubenswrapper[4808]: I0121 18:00:11.381021 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:11 crc kubenswrapper[4808]: I0121 18:00:11.381446 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:11 crc kubenswrapper[4808]: I0121 18:00:11.515958 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 21 18:00:11 crc kubenswrapper[4808]: I0121 18:00:11.516030 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:00:11 crc kubenswrapper[4808]: I0121 18:00:11.516047 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 21 18:00:11 crc kubenswrapper[4808]: I0121 18:00:11.516093 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:00:11 crc kubenswrapper[4808]: I0121 18:00:11.516151 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 21 18:00:11 crc kubenswrapper[4808]: I0121 18:00:11.516296 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:00:11 crc kubenswrapper[4808]: I0121 18:00:11.516723 4808 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:11 crc kubenswrapper[4808]: I0121 18:00:11.516754 4808 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:11 crc kubenswrapper[4808]: I0121 18:00:11.516767 4808 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:11 crc kubenswrapper[4808]: I0121 18:00:11.527277 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Jan 21 18:00:11 crc kubenswrapper[4808]: E0121 18:00:11.577236 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.102:6443: connect: connection refused" interval="400ms" Jan 21 18:00:11 crc kubenswrapper[4808]: E0121 18:00:11.978881 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.102:6443: connect: connection refused" interval="800ms" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.000405 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mzpzk" event={"ID":"688d7727-51ad-456f-a651-cea655ee7507","Type":"ContainerStarted","Data":"6e6bdbaac9a593b1edd015fb8435881eca680eb9abe413a398fb8f77fcba60e4"} Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.001039 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.001398 4808 status_manager.go:851] "Failed to get status for pod" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" pod="openshift-marketplace/redhat-marketplace-lx72r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-lx72r\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.001616 4808 status_manager.go:851] "Failed to get status for pod" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/collect-profiles-29483640-bx86q\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.001837 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.002134 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.006881 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.007868 4808 scope.go:117] "RemoveContainer" containerID="121bb0157df3897195e24f15682635bb4fd4ff48120068844746d75fc61f8e51" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.007891 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.008708 4808 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.009222 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.009672 4808 status_manager.go:851] "Failed to get status for pod" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/collect-profiles-29483640-bx86q\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.010130 4808 status_manager.go:851] "Failed to get status for pod" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" pod="openshift-marketplace/redhat-marketplace-lx72r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-lx72r\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.010408 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.010600 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.011005 4808 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.011362 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.011567 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lx72r" event={"ID":"b05e5192-b448-446a-96ba-3ec53d3644bf","Type":"ContainerStarted","Data":"02be607e2450f7f343306b4491c2582fa28e9942e7331e122cf44d05a00d02a1"} Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.011614 4808 status_manager.go:851] "Failed to get status for pod" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/collect-profiles-29483640-bx86q\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.011885 4808 status_manager.go:851] "Failed to get status for pod" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" pod="openshift-marketplace/redhat-marketplace-lx72r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-lx72r\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.012079 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.012299 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.012519 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.012687 4808 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.012851 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.013034 4808 status_manager.go:851] "Failed to get status for pod" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/collect-profiles-29483640-bx86q\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.013196 4808 status_manager.go:851] "Failed to get status for pod" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" pod="openshift-marketplace/redhat-marketplace-lx72r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-lx72r\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.013364 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.028915 4808 scope.go:117] "RemoveContainer" containerID="9ac66f79ea0501fa0e9262d92f3fd3e8851974aa0dc020577397deaf36f5a4b7" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.095006 4808 scope.go:117] "RemoveContainer" containerID="ffeecfa7e221cd92301f9990e4c2b140d53312e102f945e464c335aee4b2eed6" Jan 21 18:00:12 crc kubenswrapper[4808]: E0121 18:00:12.105232 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T18:00:12Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T18:00:12Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T18:00:12Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T18:00:12Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:aad5e438ec868272540a84dfc53b266c8a08267bec7a7617871dddeb1511dcb2\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:dd1e95af8b913ea8f010fa96cba36f2e7e5b1edfbf758c69b8c9eeb88c6911ea\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1202744046},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:2b72e40c5d5b36b681f40c16ebf3dcac6520ed0c79f174ba87f673ab7afd209a\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:d83ee77ad07e06451a84205ac4c85c69e912a1c975e1a8a95095d79218028dce\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1178956511},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:c10fecd0ba9b4f4f77af571afe82506201ee1139d1904e61b94987e47659a271\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:c44546b94a5203c84127195a969fe508a3c8e632c14d08b60a6cc3f15d19cc0d\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1167523055},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: E0121 18:00:12.106786 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: E0121 18:00:12.107110 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: E0121 18:00:12.107331 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: E0121 18:00:12.107519 4808 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:12 crc kubenswrapper[4808]: E0121 18:00:12.107546 4808 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.113996 4808 scope.go:117] "RemoveContainer" containerID="93390173555e6c2a055307b1ea5c8893de0d142516b82bec031bddc0fb5f31a1" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.134313 4808 scope.go:117] "RemoveContainer" containerID="b5213509f2f53fab3372f2b9874f0fc484bbd490bbead89a66676b2f148d9159" Jan 21 18:00:12 crc kubenswrapper[4808]: I0121 18:00:12.162640 4808 scope.go:117] "RemoveContainer" containerID="c69e06781dbe60d50e118f6fd4868bdcbf9e6eaabd886fd544be7412f6b4f9f1" Jan 21 18:00:12 crc kubenswrapper[4808]: E0121 18:00:12.780046 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.102:6443: connect: connection refused" interval="1.6s" Jan 21 18:00:13 crc kubenswrapper[4808]: I0121 18:00:13.018082 4808 generic.go:334] "Generic (PLEG): container finished" podID="688d7727-51ad-456f-a651-cea655ee7507" containerID="6e6bdbaac9a593b1edd015fb8435881eca680eb9abe413a398fb8f77fcba60e4" exitCode=0 Jan 21 18:00:13 crc kubenswrapper[4808]: I0121 18:00:13.018376 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mzpzk" event={"ID":"688d7727-51ad-456f-a651-cea655ee7507","Type":"ContainerDied","Data":"6e6bdbaac9a593b1edd015fb8435881eca680eb9abe413a398fb8f77fcba60e4"} Jan 21 18:00:13 crc kubenswrapper[4808]: I0121 18:00:13.019120 4808 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:13 crc kubenswrapper[4808]: I0121 18:00:13.019418 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:13 crc kubenswrapper[4808]: I0121 18:00:13.019809 4808 status_manager.go:851] "Failed to get status for pod" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/collect-profiles-29483640-bx86q\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:13 crc kubenswrapper[4808]: I0121 18:00:13.020595 4808 status_manager.go:851] "Failed to get status for pod" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" pod="openshift-marketplace/redhat-marketplace-lx72r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-lx72r\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:13 crc kubenswrapper[4808]: I0121 18:00:13.021140 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:13 crc kubenswrapper[4808]: I0121 18:00:13.021432 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:14 crc kubenswrapper[4808]: E0121 18:00:14.381285 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.102:6443: connect: connection refused" interval="3.2s" Jan 21 18:00:14 crc kubenswrapper[4808]: E0121 18:00:14.429883 4808 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.102:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188cd0de77a21d78 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-21 18:00:07.558192504 +0000 UTC m=+242.288703399,LastTimestamp:2026-01-21 18:00:07.558192504 +0000 UTC m=+242.288703399,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 21 18:00:14 crc kubenswrapper[4808]: I0121 18:00:14.689264 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lx72r" Jan 21 18:00:14 crc kubenswrapper[4808]: I0121 18:00:14.689579 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lx72r" Jan 21 18:00:14 crc kubenswrapper[4808]: I0121 18:00:14.759531 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lx72r" Jan 21 18:00:14 crc kubenswrapper[4808]: I0121 18:00:14.760358 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:14 crc kubenswrapper[4808]: I0121 18:00:14.761368 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:14 crc kubenswrapper[4808]: I0121 18:00:14.762083 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:14 crc kubenswrapper[4808]: I0121 18:00:14.762394 4808 status_manager.go:851] "Failed to get status for pod" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/collect-profiles-29483640-bx86q\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:14 crc kubenswrapper[4808]: I0121 18:00:14.762807 4808 status_manager.go:851] "Failed to get status for pod" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" pod="openshift-marketplace/redhat-marketplace-lx72r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-lx72r\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:15 crc kubenswrapper[4808]: I0121 18:00:15.527493 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:15 crc kubenswrapper[4808]: I0121 18:00:15.528724 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:15 crc kubenswrapper[4808]: I0121 18:00:15.529384 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:15 crc kubenswrapper[4808]: I0121 18:00:15.529612 4808 status_manager.go:851] "Failed to get status for pod" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/collect-profiles-29483640-bx86q\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:15 crc kubenswrapper[4808]: I0121 18:00:15.529794 4808 status_manager.go:851] "Failed to get status for pod" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" pod="openshift-marketplace/redhat-marketplace-lx72r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-lx72r\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:16 crc kubenswrapper[4808]: I0121 18:00:16.044411 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mzpzk" event={"ID":"688d7727-51ad-456f-a651-cea655ee7507","Type":"ContainerStarted","Data":"17fcc91cef4afbff98c9974d450d2893c4a4892d26b05cc5074d082c4a536f4e"} Jan 21 18:00:16 crc kubenswrapper[4808]: I0121 18:00:16.045264 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:16 crc kubenswrapper[4808]: I0121 18:00:16.046498 4808 status_manager.go:851] "Failed to get status for pod" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/collect-profiles-29483640-bx86q\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:16 crc kubenswrapper[4808]: I0121 18:00:16.046844 4808 status_manager.go:851] "Failed to get status for pod" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" pod="openshift-marketplace/redhat-marketplace-lx72r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-lx72r\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:16 crc kubenswrapper[4808]: I0121 18:00:16.047027 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:16 crc kubenswrapper[4808]: I0121 18:00:16.047199 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:16 crc kubenswrapper[4808]: I0121 18:00:16.080891 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lx72r" Jan 21 18:00:16 crc kubenswrapper[4808]: I0121 18:00:16.081813 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:16 crc kubenswrapper[4808]: I0121 18:00:16.082185 4808 status_manager.go:851] "Failed to get status for pod" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" pod="openshift-marketplace/redhat-marketplace-lx72r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-lx72r\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:16 crc kubenswrapper[4808]: I0121 18:00:16.082543 4808 status_manager.go:851] "Failed to get status for pod" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/collect-profiles-29483640-bx86q\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:16 crc kubenswrapper[4808]: I0121 18:00:16.082892 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:16 crc kubenswrapper[4808]: I0121 18:00:16.083219 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:17 crc kubenswrapper[4808]: I0121 18:00:17.080776 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mzpzk" Jan 21 18:00:17 crc kubenswrapper[4808]: I0121 18:00:17.081520 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mzpzk" Jan 21 18:00:17 crc kubenswrapper[4808]: E0121 18:00:17.582967 4808 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.102:6443: connect: connection refused" interval="6.4s" Jan 21 18:00:18 crc kubenswrapper[4808]: I0121 18:00:18.126199 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-mzpzk" podUID="688d7727-51ad-456f-a651-cea655ee7507" containerName="registry-server" probeResult="failure" output=< Jan 21 18:00:18 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Jan 21 18:00:18 crc kubenswrapper[4808]: > Jan 21 18:00:18 crc kubenswrapper[4808]: I0121 18:00:18.520153 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:18 crc kubenswrapper[4808]: I0121 18:00:18.521367 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:18 crc kubenswrapper[4808]: I0121 18:00:18.521911 4808 status_manager.go:851] "Failed to get status for pod" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/collect-profiles-29483640-bx86q\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:18 crc kubenswrapper[4808]: I0121 18:00:18.522343 4808 status_manager.go:851] "Failed to get status for pod" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" pod="openshift-marketplace/redhat-marketplace-lx72r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-lx72r\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:18 crc kubenswrapper[4808]: I0121 18:00:18.522768 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:18 crc kubenswrapper[4808]: I0121 18:00:18.523206 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:18 crc kubenswrapper[4808]: I0121 18:00:18.545130 4808 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1bac69ca-5e64-4274-b84e-5fcb32a3e63b" Jan 21 18:00:18 crc kubenswrapper[4808]: I0121 18:00:18.545178 4808 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1bac69ca-5e64-4274-b84e-5fcb32a3e63b" Jan 21 18:00:18 crc kubenswrapper[4808]: E0121 18:00:18.545768 4808 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:18 crc kubenswrapper[4808]: I0121 18:00:18.546521 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:18 crc kubenswrapper[4808]: W0121 18:00:18.582956 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-01972740b34b9079f1512507f36c90ac62918053488b1933de0bbcab11ce7d0f WatchSource:0}: Error finding container 01972740b34b9079f1512507f36c90ac62918053488b1933de0bbcab11ce7d0f: Status 404 returned error can't find the container with id 01972740b34b9079f1512507f36c90ac62918053488b1933de0bbcab11ce7d0f Jan 21 18:00:19 crc kubenswrapper[4808]: I0121 18:00:19.061510 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"01972740b34b9079f1512507f36c90ac62918053488b1933de0bbcab11ce7d0f"} Jan 21 18:00:20 crc kubenswrapper[4808]: I0121 18:00:20.014804 4808 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": read tcp 192.168.126.11:55394->192.168.126.11:10257: read: connection reset by peer" start-of-body= Jan 21 18:00:20 crc kubenswrapper[4808]: I0121 18:00:20.014952 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": read tcp 192.168.126.11:55394->192.168.126.11:10257: read: connection reset by peer" Jan 21 18:00:20 crc kubenswrapper[4808]: I0121 18:00:20.074926 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7e3823317236c2a95d593485d7f744f935aa9f31d9143a733bef16e53b56c4ce"} Jan 21 18:00:20 crc kubenswrapper[4808]: I0121 18:00:20.075638 4808 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1bac69ca-5e64-4274-b84e-5fcb32a3e63b" Jan 21 18:00:20 crc kubenswrapper[4808]: I0121 18:00:20.075682 4808 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1bac69ca-5e64-4274-b84e-5fcb32a3e63b" Jan 21 18:00:20 crc kubenswrapper[4808]: I0121 18:00:20.075955 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:20 crc kubenswrapper[4808]: E0121 18:00:20.076400 4808 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:20 crc kubenswrapper[4808]: I0121 18:00:20.076585 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:20 crc kubenswrapper[4808]: I0121 18:00:20.077047 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:20 crc kubenswrapper[4808]: I0121 18:00:20.077408 4808 status_manager.go:851] "Failed to get status for pod" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/collect-profiles-29483640-bx86q\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:20 crc kubenswrapper[4808]: I0121 18:00:20.077853 4808 status_manager.go:851] "Failed to get status for pod" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" pod="openshift-marketplace/redhat-marketplace-lx72r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-lx72r\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:20 crc kubenswrapper[4808]: I0121 18:00:20.222739 4808 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 21 18:00:20 crc kubenswrapper[4808]: I0121 18:00:20.222815 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 21 18:00:21 crc kubenswrapper[4808]: I0121 18:00:21.088586 4808 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="7e3823317236c2a95d593485d7f744f935aa9f31d9143a733bef16e53b56c4ce" exitCode=0 Jan 21 18:00:21 crc kubenswrapper[4808]: I0121 18:00:21.088777 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"7e3823317236c2a95d593485d7f744f935aa9f31d9143a733bef16e53b56c4ce"} Jan 21 18:00:21 crc kubenswrapper[4808]: I0121 18:00:21.089303 4808 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1bac69ca-5e64-4274-b84e-5fcb32a3e63b" Jan 21 18:00:21 crc kubenswrapper[4808]: I0121 18:00:21.089346 4808 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1bac69ca-5e64-4274-b84e-5fcb32a3e63b" Jan 21 18:00:21 crc kubenswrapper[4808]: E0121 18:00:21.089671 4808 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:21 crc kubenswrapper[4808]: I0121 18:00:21.089677 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:21 crc kubenswrapper[4808]: I0121 18:00:21.090140 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:21 crc kubenswrapper[4808]: I0121 18:00:21.090906 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:21 crc kubenswrapper[4808]: I0121 18:00:21.091421 4808 status_manager.go:851] "Failed to get status for pod" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/collect-profiles-29483640-bx86q\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:21 crc kubenswrapper[4808]: I0121 18:00:21.091843 4808 status_manager.go:851] "Failed to get status for pod" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" pod="openshift-marketplace/redhat-marketplace-lx72r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-lx72r\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:21 crc kubenswrapper[4808]: I0121 18:00:21.095766 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 21 18:00:21 crc kubenswrapper[4808]: I0121 18:00:21.095822 4808 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92" exitCode=1 Jan 21 18:00:21 crc kubenswrapper[4808]: I0121 18:00:21.095850 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92"} Jan 21 18:00:21 crc kubenswrapper[4808]: I0121 18:00:21.096259 4808 scope.go:117] "RemoveContainer" containerID="3ce78ed12702c1ec09f8eee415a35a023f90c1ae51e155e161e01f716b3d3b92" Jan 21 18:00:21 crc kubenswrapper[4808]: I0121 18:00:21.096931 4808 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:21 crc kubenswrapper[4808]: I0121 18:00:21.097520 4808 status_manager.go:851] "Failed to get status for pod" podUID="688d7727-51ad-456f-a651-cea655ee7507" pod="openshift-marketplace/community-operators-mzpzk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mzpzk\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:21 crc kubenswrapper[4808]: I0121 18:00:21.098099 4808 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:21 crc kubenswrapper[4808]: I0121 18:00:21.098519 4808 status_manager.go:851] "Failed to get status for pod" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:21 crc kubenswrapper[4808]: I0121 18:00:21.098852 4808 status_manager.go:851] "Failed to get status for pod" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" pod="openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/collect-profiles-29483640-bx86q\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:21 crc kubenswrapper[4808]: I0121 18:00:21.099198 4808 status_manager.go:851] "Failed to get status for pod" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" pod="openshift-marketplace/redhat-marketplace-lx72r" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-lx72r\": dial tcp 38.129.56.102:6443: connect: connection refused" Jan 21 18:00:22 crc kubenswrapper[4808]: I0121 18:00:22.103024 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"cb9348e4be65ddac7d08e719f1bbb6dd6a7c3a9414a1d55f352714604da101ff"} Jan 21 18:00:22 crc kubenswrapper[4808]: I0121 18:00:22.103738 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6c1bb8621e6fc6fb90e1a30aef79be77f30c8b829281d4d9d6b938e6c07cd958"} Jan 21 18:00:22 crc kubenswrapper[4808]: I0121 18:00:22.105770 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 21 18:00:22 crc kubenswrapper[4808]: I0121 18:00:22.105831 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c9a7e1bdbb7091328184a34ccfea80a7c4583b4a107206d9fca593e42c0c4409"} Jan 21 18:00:23 crc kubenswrapper[4808]: I0121 18:00:23.116324 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"eb309c8bfb4457295df3cccce9c7040a39864c9c6e0143ab62920e255070e133"} Jan 21 18:00:23 crc kubenswrapper[4808]: I0121 18:00:23.116379 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8ecb915d83562b9d6f204d460144058445804780d0c25237f549fe3aeeb28cd5"} Jan 21 18:00:24 crc kubenswrapper[4808]: I0121 18:00:24.050076 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-5v5c4" Jan 21 18:00:24 crc kubenswrapper[4808]: I0121 18:00:24.125424 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"dc8c65dc1ffbe7bc50df620c66233b0cdb0ef8068bbcf5590f645bea11656d2c"} Jan 21 18:00:24 crc kubenswrapper[4808]: I0121 18:00:24.125626 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:24 crc kubenswrapper[4808]: I0121 18:00:24.125758 4808 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1bac69ca-5e64-4274-b84e-5fcb32a3e63b" Jan 21 18:00:24 crc kubenswrapper[4808]: I0121 18:00:24.125793 4808 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1bac69ca-5e64-4274-b84e-5fcb32a3e63b" Jan 21 18:00:26 crc kubenswrapper[4808]: I0121 18:00:26.784105 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" podUID="66c5c3c8-4515-401a-bfe9-5755068bc771" containerName="oauth-openshift" containerID="cri-o://c566ddd31fa616ed33feec6e779ffa61917579a6a973416a023738d2e73a78a4" gracePeriod=15 Jan 21 18:00:27 crc kubenswrapper[4808]: I0121 18:00:27.128770 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mzpzk" Jan 21 18:00:27 crc kubenswrapper[4808]: I0121 18:00:27.134263 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 18:00:27 crc kubenswrapper[4808]: I0121 18:00:27.143417 4808 generic.go:334] "Generic (PLEG): container finished" podID="66c5c3c8-4515-401a-bfe9-5755068bc771" containerID="c566ddd31fa616ed33feec6e779ffa61917579a6a973416a023738d2e73a78a4" exitCode=0 Jan 21 18:00:27 crc kubenswrapper[4808]: I0121 18:00:27.143458 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" event={"ID":"66c5c3c8-4515-401a-bfe9-5755068bc771","Type":"ContainerDied","Data":"c566ddd31fa616ed33feec6e779ffa61917579a6a973416a023738d2e73a78a4"} Jan 21 18:00:27 crc kubenswrapper[4808]: I0121 18:00:27.143485 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" event={"ID":"66c5c3c8-4515-401a-bfe9-5755068bc771","Type":"ContainerDied","Data":"af2baea7f35d573815130e53461594fc2a9014abe94f43e837d136de3413e5a8"} Jan 21 18:00:27 crc kubenswrapper[4808]: I0121 18:00:27.143501 4808 scope.go:117] "RemoveContainer" containerID="c566ddd31fa616ed33feec6e779ffa61917579a6a973416a023738d2e73a78a4" Jan 21 18:00:27 crc kubenswrapper[4808]: I0121 18:00:27.143521 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bz45s" Jan 21 18:00:27 crc kubenswrapper[4808]: I0121 18:00:27.167516 4808 scope.go:117] "RemoveContainer" containerID="c566ddd31fa616ed33feec6e779ffa61917579a6a973416a023738d2e73a78a4" Jan 21 18:00:27 crc kubenswrapper[4808]: E0121 18:00:27.167944 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c566ddd31fa616ed33feec6e779ffa61917579a6a973416a023738d2e73a78a4\": container with ID starting with c566ddd31fa616ed33feec6e779ffa61917579a6a973416a023738d2e73a78a4 not found: ID does not exist" containerID="c566ddd31fa616ed33feec6e779ffa61917579a6a973416a023738d2e73a78a4" Jan 21 18:00:27 crc kubenswrapper[4808]: I0121 18:00:27.167981 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c566ddd31fa616ed33feec6e779ffa61917579a6a973416a023738d2e73a78a4"} err="failed to get container status \"c566ddd31fa616ed33feec6e779ffa61917579a6a973416a023738d2e73a78a4\": rpc error: code = NotFound desc = could not find container \"c566ddd31fa616ed33feec6e779ffa61917579a6a973416a023738d2e73a78a4\": container with ID starting with c566ddd31fa616ed33feec6e779ffa61917579a6a973416a023738d2e73a78a4 not found: ID does not exist" Jan 21 18:00:27 crc kubenswrapper[4808]: I0121 18:00:27.179154 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mzpzk" Jan 21 18:00:28 crc kubenswrapper[4808]: I0121 18:00:28.548410 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:28 crc kubenswrapper[4808]: I0121 18:00:28.549735 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:28 crc kubenswrapper[4808]: I0121 18:00:28.554863 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.116898 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.116950 4808 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.116994 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.134678 4808 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.165171 4808 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1bac69ca-5e64-4274-b84e-5fcb32a3e63b" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.165264 4808 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1bac69ca-5e64-4274-b84e-5fcb32a3e63b" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.168850 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.171181 4808 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="e10dffe0-e3dc-4609-b200-b0158304247f" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.172469 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-ocp-branding-template\") pod \"66c5c3c8-4515-401a-bfe9-5755068bc771\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.172528 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-audit-policies\") pod \"66c5c3c8-4515-401a-bfe9-5755068bc771\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.172562 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-template-provider-selection\") pod \"66c5c3c8-4515-401a-bfe9-5755068bc771\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.172639 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-idp-0-file-data\") pod \"66c5c3c8-4515-401a-bfe9-5755068bc771\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.172686 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-template-error\") pod \"66c5c3c8-4515-401a-bfe9-5755068bc771\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.172709 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-service-ca\") pod \"66c5c3c8-4515-401a-bfe9-5755068bc771\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.172764 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-trusted-ca-bundle\") pod \"66c5c3c8-4515-401a-bfe9-5755068bc771\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.172785 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66c5c3c8-4515-401a-bfe9-5755068bc771-audit-dir\") pod \"66c5c3c8-4515-401a-bfe9-5755068bc771\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.172809 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-session\") pod \"66c5c3c8-4515-401a-bfe9-5755068bc771\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.172827 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-serving-cert\") pod \"66c5c3c8-4515-401a-bfe9-5755068bc771\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.172843 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-template-login\") pod \"66c5c3c8-4515-401a-bfe9-5755068bc771\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.172871 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skpzt\" (UniqueName: \"kubernetes.io/projected/66c5c3c8-4515-401a-bfe9-5755068bc771-kube-api-access-skpzt\") pod \"66c5c3c8-4515-401a-bfe9-5755068bc771\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.172897 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-router-certs\") pod \"66c5c3c8-4515-401a-bfe9-5755068bc771\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.172913 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-cliconfig\") pod \"66c5c3c8-4515-401a-bfe9-5755068bc771\" (UID: \"66c5c3c8-4515-401a-bfe9-5755068bc771\") " Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.174639 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "66c5c3c8-4515-401a-bfe9-5755068bc771" (UID: "66c5c3c8-4515-401a-bfe9-5755068bc771"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.175180 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "66c5c3c8-4515-401a-bfe9-5755068bc771" (UID: "66c5c3c8-4515-401a-bfe9-5755068bc771"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.177284 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66c5c3c8-4515-401a-bfe9-5755068bc771-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "66c5c3c8-4515-401a-bfe9-5755068bc771" (UID: "66c5c3c8-4515-401a-bfe9-5755068bc771"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.181956 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "66c5c3c8-4515-401a-bfe9-5755068bc771" (UID: "66c5c3c8-4515-401a-bfe9-5755068bc771"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.182486 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "66c5c3c8-4515-401a-bfe9-5755068bc771" (UID: "66c5c3c8-4515-401a-bfe9-5755068bc771"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.186187 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66c5c3c8-4515-401a-bfe9-5755068bc771-kube-api-access-skpzt" (OuterVolumeSpecName: "kube-api-access-skpzt") pod "66c5c3c8-4515-401a-bfe9-5755068bc771" (UID: "66c5c3c8-4515-401a-bfe9-5755068bc771"). InnerVolumeSpecName "kube-api-access-skpzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.197567 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "66c5c3c8-4515-401a-bfe9-5755068bc771" (UID: "66c5c3c8-4515-401a-bfe9-5755068bc771"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.198868 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "66c5c3c8-4515-401a-bfe9-5755068bc771" (UID: "66c5c3c8-4515-401a-bfe9-5755068bc771"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.199718 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "66c5c3c8-4515-401a-bfe9-5755068bc771" (UID: "66c5c3c8-4515-401a-bfe9-5755068bc771"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.200081 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "66c5c3c8-4515-401a-bfe9-5755068bc771" (UID: "66c5c3c8-4515-401a-bfe9-5755068bc771"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.206001 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "66c5c3c8-4515-401a-bfe9-5755068bc771" (UID: "66c5c3c8-4515-401a-bfe9-5755068bc771"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.206127 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "66c5c3c8-4515-401a-bfe9-5755068bc771" (UID: "66c5c3c8-4515-401a-bfe9-5755068bc771"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.206278 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "66c5c3c8-4515-401a-bfe9-5755068bc771" (UID: "66c5c3c8-4515-401a-bfe9-5755068bc771"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.206967 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "66c5c3c8-4515-401a-bfe9-5755068bc771" (UID: "66c5c3c8-4515-401a-bfe9-5755068bc771"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.274548 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.275149 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.275178 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.275196 4808 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.275216 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.275235 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.275276 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.275293 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.275311 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.275329 4808 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66c5c3c8-4515-401a-bfe9-5755068bc771-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.275346 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.275363 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.275379 4808 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66c5c3c8-4515-401a-bfe9-5755068bc771-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:29 crc kubenswrapper[4808]: I0121 18:00:29.275397 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skpzt\" (UniqueName: \"kubernetes.io/projected/66c5c3c8-4515-401a-bfe9-5755068bc771-kube-api-access-skpzt\") on node \"crc\" DevicePath \"\"" Jan 21 18:00:30 crc kubenswrapper[4808]: I0121 18:00:30.012431 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 18:00:30 crc kubenswrapper[4808]: I0121 18:00:30.169908 4808 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1bac69ca-5e64-4274-b84e-5fcb32a3e63b" Jan 21 18:00:30 crc kubenswrapper[4808]: I0121 18:00:30.169953 4808 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1bac69ca-5e64-4274-b84e-5fcb32a3e63b" Jan 21 18:00:35 crc kubenswrapper[4808]: I0121 18:00:35.546745 4808 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="e10dffe0-e3dc-4609-b200-b0158304247f" Jan 21 18:00:38 crc kubenswrapper[4808]: I0121 18:00:38.343758 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 21 18:00:38 crc kubenswrapper[4808]: I0121 18:00:38.595432 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 21 18:00:38 crc kubenswrapper[4808]: I0121 18:00:38.730399 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 21 18:00:38 crc kubenswrapper[4808]: I0121 18:00:38.932013 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 21 18:00:38 crc kubenswrapper[4808]: I0121 18:00:38.934576 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 21 18:00:39 crc kubenswrapper[4808]: I0121 18:00:39.121870 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 18:00:39 crc kubenswrapper[4808]: I0121 18:00:39.126883 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 18:00:39 crc kubenswrapper[4808]: I0121 18:00:39.238943 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 21 18:00:39 crc kubenswrapper[4808]: I0121 18:00:39.240416 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 21 18:00:39 crc kubenswrapper[4808]: I0121 18:00:39.766026 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 21 18:00:39 crc kubenswrapper[4808]: I0121 18:00:39.892812 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 21 18:00:39 crc kubenswrapper[4808]: I0121 18:00:39.982923 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 21 18:00:39 crc kubenswrapper[4808]: I0121 18:00:39.994621 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 21 18:00:40 crc kubenswrapper[4808]: I0121 18:00:40.058091 4808 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 21 18:00:40 crc kubenswrapper[4808]: I0121 18:00:40.068352 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 21 18:00:40 crc kubenswrapper[4808]: I0121 18:00:40.404600 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 21 18:00:40 crc kubenswrapper[4808]: I0121 18:00:40.466540 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 21 18:00:40 crc kubenswrapper[4808]: I0121 18:00:40.716593 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 21 18:00:40 crc kubenswrapper[4808]: I0121 18:00:40.816004 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 21 18:00:40 crc kubenswrapper[4808]: I0121 18:00:40.894845 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 21 18:00:41 crc kubenswrapper[4808]: I0121 18:00:41.223942 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 21 18:00:41 crc kubenswrapper[4808]: I0121 18:00:41.227822 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 21 18:00:41 crc kubenswrapper[4808]: I0121 18:00:41.389376 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 21 18:00:41 crc kubenswrapper[4808]: I0121 18:00:41.401551 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 21 18:00:41 crc kubenswrapper[4808]: I0121 18:00:41.404027 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 21 18:00:41 crc kubenswrapper[4808]: I0121 18:00:41.410919 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 21 18:00:41 crc kubenswrapper[4808]: I0121 18:00:41.534811 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 21 18:00:41 crc kubenswrapper[4808]: I0121 18:00:41.763498 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 21 18:00:41 crc kubenswrapper[4808]: I0121 18:00:41.778570 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 21 18:00:41 crc kubenswrapper[4808]: I0121 18:00:41.821528 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 21 18:00:41 crc kubenswrapper[4808]: I0121 18:00:41.866135 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 21 18:00:41 crc kubenswrapper[4808]: I0121 18:00:41.973912 4808 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 21 18:00:42 crc kubenswrapper[4808]: I0121 18:00:42.053166 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 21 18:00:42 crc kubenswrapper[4808]: I0121 18:00:42.245016 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 21 18:00:42 crc kubenswrapper[4808]: I0121 18:00:42.259198 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 21 18:00:42 crc kubenswrapper[4808]: I0121 18:00:42.372047 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 21 18:00:42 crc kubenswrapper[4808]: I0121 18:00:42.493170 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 21 18:00:42 crc kubenswrapper[4808]: I0121 18:00:42.505375 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 21 18:00:42 crc kubenswrapper[4808]: I0121 18:00:42.536983 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 21 18:00:42 crc kubenswrapper[4808]: I0121 18:00:42.556617 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 21 18:00:42 crc kubenswrapper[4808]: I0121 18:00:42.815910 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 21 18:00:42 crc kubenswrapper[4808]: I0121 18:00:42.842862 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 21 18:00:42 crc kubenswrapper[4808]: I0121 18:00:42.923114 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 21 18:00:42 crc kubenswrapper[4808]: I0121 18:00:42.951237 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 21 18:00:43 crc kubenswrapper[4808]: I0121 18:00:43.046489 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 21 18:00:43 crc kubenswrapper[4808]: I0121 18:00:43.058137 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 21 18:00:43 crc kubenswrapper[4808]: I0121 18:00:43.076435 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 21 18:00:43 crc kubenswrapper[4808]: I0121 18:00:43.099565 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 21 18:00:43 crc kubenswrapper[4808]: I0121 18:00:43.119225 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 21 18:00:43 crc kubenswrapper[4808]: I0121 18:00:43.120065 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 21 18:00:43 crc kubenswrapper[4808]: I0121 18:00:43.128797 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 21 18:00:43 crc kubenswrapper[4808]: I0121 18:00:43.257975 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 21 18:00:43 crc kubenswrapper[4808]: I0121 18:00:43.326559 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 21 18:00:43 crc kubenswrapper[4808]: I0121 18:00:43.337804 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 21 18:00:43 crc kubenswrapper[4808]: I0121 18:00:43.353184 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 21 18:00:43 crc kubenswrapper[4808]: I0121 18:00:43.363234 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 21 18:00:43 crc kubenswrapper[4808]: I0121 18:00:43.592409 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 21 18:00:43 crc kubenswrapper[4808]: I0121 18:00:43.643136 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 21 18:00:43 crc kubenswrapper[4808]: I0121 18:00:43.803383 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 21 18:00:43 crc kubenswrapper[4808]: I0121 18:00:43.837191 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 21 18:00:43 crc kubenswrapper[4808]: I0121 18:00:43.959132 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 21 18:00:43 crc kubenswrapper[4808]: I0121 18:00:43.964456 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 21 18:00:43 crc kubenswrapper[4808]: I0121 18:00:43.988834 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 21 18:00:43 crc kubenswrapper[4808]: I0121 18:00:43.991119 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 21 18:00:44 crc kubenswrapper[4808]: I0121 18:00:44.017014 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 21 18:00:44 crc kubenswrapper[4808]: I0121 18:00:44.066278 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 21 18:00:44 crc kubenswrapper[4808]: I0121 18:00:44.111383 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 21 18:00:44 crc kubenswrapper[4808]: I0121 18:00:44.150612 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 21 18:00:44 crc kubenswrapper[4808]: I0121 18:00:44.157064 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 21 18:00:44 crc kubenswrapper[4808]: I0121 18:00:44.158605 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 21 18:00:44 crc kubenswrapper[4808]: I0121 18:00:44.159892 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 21 18:00:44 crc kubenswrapper[4808]: I0121 18:00:44.204492 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 21 18:00:44 crc kubenswrapper[4808]: I0121 18:00:44.273316 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 21 18:00:44 crc kubenswrapper[4808]: I0121 18:00:44.312035 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 21 18:00:44 crc kubenswrapper[4808]: I0121 18:00:44.396658 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 21 18:00:44 crc kubenswrapper[4808]: I0121 18:00:44.586085 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 21 18:00:44 crc kubenswrapper[4808]: I0121 18:00:44.588117 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 21 18:00:44 crc kubenswrapper[4808]: I0121 18:00:44.597267 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 21 18:00:44 crc kubenswrapper[4808]: I0121 18:00:44.645388 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 21 18:00:44 crc kubenswrapper[4808]: I0121 18:00:44.718480 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 21 18:00:44 crc kubenswrapper[4808]: I0121 18:00:44.902463 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 21 18:00:44 crc kubenswrapper[4808]: I0121 18:00:44.962986 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 21 18:00:44 crc kubenswrapper[4808]: I0121 18:00:44.987158 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.031757 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.042002 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.102168 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.141395 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.210100 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.384723 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.459148 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.477847 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.496515 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.513167 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.516868 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.534898 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.622448 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.667049 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.753159 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.794517 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.795777 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.838572 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.874184 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.963696 4808 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 21 18:00:45 crc kubenswrapper[4808]: I0121 18:00:45.973350 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 21 18:00:46 crc kubenswrapper[4808]: I0121 18:00:46.008397 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 21 18:00:46 crc kubenswrapper[4808]: I0121 18:00:46.023368 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 21 18:00:46 crc kubenswrapper[4808]: I0121 18:00:46.094561 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 21 18:00:46 crc kubenswrapper[4808]: I0121 18:00:46.102466 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 21 18:00:46 crc kubenswrapper[4808]: I0121 18:00:46.199896 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 21 18:00:46 crc kubenswrapper[4808]: I0121 18:00:46.208404 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 21 18:00:46 crc kubenswrapper[4808]: I0121 18:00:46.222117 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 21 18:00:46 crc kubenswrapper[4808]: I0121 18:00:46.356749 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 21 18:00:46 crc kubenswrapper[4808]: I0121 18:00:46.358963 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 21 18:00:46 crc kubenswrapper[4808]: I0121 18:00:46.434053 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 21 18:00:46 crc kubenswrapper[4808]: I0121 18:00:46.435807 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 21 18:00:46 crc kubenswrapper[4808]: I0121 18:00:46.554796 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 21 18:00:46 crc kubenswrapper[4808]: I0121 18:00:46.585821 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 21 18:00:46 crc kubenswrapper[4808]: I0121 18:00:46.608500 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 21 18:00:46 crc kubenswrapper[4808]: I0121 18:00:46.704460 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 21 18:00:46 crc kubenswrapper[4808]: I0121 18:00:46.763403 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 21 18:00:46 crc kubenswrapper[4808]: I0121 18:00:46.791179 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 21 18:00:46 crc kubenswrapper[4808]: I0121 18:00:46.855741 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 21 18:00:47 crc kubenswrapper[4808]: I0121 18:00:47.022308 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 21 18:00:47 crc kubenswrapper[4808]: I0121 18:00:47.048378 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 21 18:00:47 crc kubenswrapper[4808]: I0121 18:00:47.060882 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 21 18:00:47 crc kubenswrapper[4808]: I0121 18:00:47.080184 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 21 18:00:47 crc kubenswrapper[4808]: I0121 18:00:47.138015 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 21 18:00:47 crc kubenswrapper[4808]: I0121 18:00:47.163813 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 21 18:00:47 crc kubenswrapper[4808]: I0121 18:00:47.250493 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 21 18:00:47 crc kubenswrapper[4808]: I0121 18:00:47.329990 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 21 18:00:47 crc kubenswrapper[4808]: I0121 18:00:47.335693 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 21 18:00:47 crc kubenswrapper[4808]: I0121 18:00:47.342646 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 21 18:00:47 crc kubenswrapper[4808]: I0121 18:00:47.447499 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 21 18:00:47 crc kubenswrapper[4808]: I0121 18:00:47.619115 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:47.631872 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:47.673235 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:47.690374 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:47.723650 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:47.736042 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:47.788505 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:47.911956 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:47.922456 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:47.949096 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:47.952828 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:47.974771 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:48.004315 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:48.222310 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:48.229304 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:48.242376 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:48.271069 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:48.324726 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:48.340505 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:48.366671 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:48.514659 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:48.670734 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:48.740510 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 21 18:00:48 crc kubenswrapper[4808]: I0121 18:00:48.756295 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 21 18:00:49 crc kubenswrapper[4808]: I0121 18:00:49.023079 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 21 18:00:49 crc kubenswrapper[4808]: I0121 18:00:49.101214 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 21 18:00:49 crc kubenswrapper[4808]: I0121 18:00:49.160884 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 21 18:00:49 crc kubenswrapper[4808]: I0121 18:00:49.213693 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 21 18:00:49 crc kubenswrapper[4808]: I0121 18:00:49.235341 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 21 18:00:49 crc kubenswrapper[4808]: I0121 18:00:49.257589 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 21 18:00:49 crc kubenswrapper[4808]: I0121 18:00:49.317038 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 21 18:00:49 crc kubenswrapper[4808]: I0121 18:00:49.354553 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 21 18:00:49 crc kubenswrapper[4808]: I0121 18:00:49.407077 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 21 18:00:49 crc kubenswrapper[4808]: I0121 18:00:49.417512 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 21 18:00:49 crc kubenswrapper[4808]: I0121 18:00:49.422043 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 21 18:00:49 crc kubenswrapper[4808]: I0121 18:00:49.486961 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 21 18:00:49 crc kubenswrapper[4808]: I0121 18:00:49.569741 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 21 18:00:49 crc kubenswrapper[4808]: I0121 18:00:49.979080 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.005162 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.028009 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.080142 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.097104 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.152311 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.167131 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.186427 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.279300 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.286336 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.331350 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.364086 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.373453 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.401995 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.403182 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.421295 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.425368 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.521306 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.622087 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.721214 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.740061 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.835977 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.952685 4808 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 21 18:00:50 crc kubenswrapper[4808]: I0121 18:00:50.956142 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.104766 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.190558 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.242794 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.251464 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.263033 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.294447 4808 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.294716 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=44.294699947 podStartE2EDuration="44.294699947s" podCreationTimestamp="2026-01-21 18:00:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 18:00:28.193578695 +0000 UTC m=+262.924089580" watchObservedRunningTime="2026-01-21 18:00:51.294699947 +0000 UTC m=+286.025210832" Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.295029 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mzpzk" podStartSLOduration=38.952052073 podStartE2EDuration="45.295022685s" podCreationTimestamp="2026-01-21 18:00:06 +0000 UTC" firstStartedPulling="2026-01-21 18:00:08.964832515 +0000 UTC m=+243.695343400" lastFinishedPulling="2026-01-21 18:00:15.307803117 +0000 UTC m=+250.038314012" observedRunningTime="2026-01-21 18:00:28.275976545 +0000 UTC m=+263.006487430" watchObservedRunningTime="2026-01-21 18:00:51.295022685 +0000 UTC m=+286.025533570" Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.297727 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lx72r" podStartSLOduration=41.728719414 podStartE2EDuration="47.297720863s" podCreationTimestamp="2026-01-21 18:00:04 +0000 UTC" firstStartedPulling="2026-01-21 18:00:05.903386681 +0000 UTC m=+240.633897586" lastFinishedPulling="2026-01-21 18:00:11.47238811 +0000 UTC m=+246.202899035" observedRunningTime="2026-01-21 18:00:28.2404982 +0000 UTC m=+262.971009095" watchObservedRunningTime="2026-01-21 18:00:51.297720863 +0000 UTC m=+286.028231738" Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.298644 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bz45s","openshift-kube-apiserver/kube-apiserver-crc"] Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.298699 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.298722 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-n56cl"] Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.303603 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.325523 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=22.325500234 podStartE2EDuration="22.325500234s" podCreationTimestamp="2026-01-21 18:00:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 18:00:51.32178116 +0000 UTC m=+286.052292045" watchObservedRunningTime="2026-01-21 18:00:51.325500234 +0000 UTC m=+286.056011119" Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.474015 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.527422 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66c5c3c8-4515-401a-bfe9-5755068bc771" path="/var/lib/kubelet/pods/66c5c3c8-4515-401a-bfe9-5755068bc771/volumes" Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.596705 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.596761 4808 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.768148 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.858309 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 21 18:00:51 crc kubenswrapper[4808]: I0121 18:00:51.930997 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.020985 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.119843 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.166643 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.169736 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.256752 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2"] Jan 21 18:00:52 crc kubenswrapper[4808]: E0121 18:00:52.256957 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" containerName="installer" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.256968 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" containerName="installer" Jan 21 18:00:52 crc kubenswrapper[4808]: E0121 18:00:52.256978 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66c5c3c8-4515-401a-bfe9-5755068bc771" containerName="oauth-openshift" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.256984 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="66c5c3c8-4515-401a-bfe9-5755068bc771" containerName="oauth-openshift" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.257075 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f152312-7210-4e7a-b9c1-6a2942f488aa" containerName="installer" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.257088 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="66c5c3c8-4515-401a-bfe9-5755068bc771" containerName="oauth-openshift" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.257651 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.260645 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.261757 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.261847 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.261974 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.262896 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.263000 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.263020 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.263036 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.263745 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.263822 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.264276 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.270330 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.271853 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.274831 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.277054 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.305277 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.317536 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.359858 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-service-ca\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.359905 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-user-template-login\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.359935 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-audit-policies\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.359999 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.360036 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-audit-dir\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.360084 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-cliconfig\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.360100 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-router-certs\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.360129 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-serving-cert\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.360164 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-session\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.360207 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.360228 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-user-template-error\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.360265 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.360281 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgttp\" (UniqueName: \"kubernetes.io/projected/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-kube-api-access-hgttp\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.360303 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.380086 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.461634 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-router-certs\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.461727 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-cliconfig\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.462525 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-serving-cert\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.462677 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-session\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.462728 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.462782 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-user-template-error\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.462843 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.462870 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgttp\" (UniqueName: \"kubernetes.io/projected/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-kube-api-access-hgttp\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.462906 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.462951 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-service-ca\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.462978 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-user-template-login\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.463015 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-audit-policies\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.463055 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-cliconfig\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.463058 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.463153 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-audit-dir\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.463307 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-audit-dir\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.463787 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-service-ca\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.463912 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.464065 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-audit-policies\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.468323 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-session\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.468338 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.468651 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-serving-cert\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.469615 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.470343 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.470556 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.471588 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-user-template-error\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.474779 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-system-router-certs\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.475127 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-v4-0-config-user-template-login\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.483961 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgttp\" (UniqueName: \"kubernetes.io/projected/8e95cdda-a4de-4569-8c2e-dd9eab1e3899-kube-api-access-hgttp\") pod \"oauth-openshift-756fdfd6b5-hjcv2\" (UID: \"8e95cdda-a4de-4569-8c2e-dd9eab1e3899\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.506522 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.578223 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.588901 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.766785 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.816514 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.850920 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.888256 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.890549 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 21 18:00:52 crc kubenswrapper[4808]: I0121 18:00:52.989524 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 21 18:00:53 crc kubenswrapper[4808]: I0121 18:00:53.188626 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 21 18:00:53 crc kubenswrapper[4808]: I0121 18:00:53.203682 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 21 18:00:53 crc kubenswrapper[4808]: I0121 18:00:53.274509 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 21 18:00:53 crc kubenswrapper[4808]: I0121 18:00:53.347492 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 21 18:00:53 crc kubenswrapper[4808]: I0121 18:00:53.362462 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2"] Jan 21 18:00:53 crc kubenswrapper[4808]: I0121 18:00:53.371708 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 21 18:00:53 crc kubenswrapper[4808]: I0121 18:00:53.409751 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 21 18:00:53 crc kubenswrapper[4808]: I0121 18:00:53.572058 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 21 18:00:53 crc kubenswrapper[4808]: I0121 18:00:53.628854 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2"] Jan 21 18:00:54 crc kubenswrapper[4808]: I0121 18:00:54.164283 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 21 18:00:54 crc kubenswrapper[4808]: I0121 18:00:54.308941 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" event={"ID":"8e95cdda-a4de-4569-8c2e-dd9eab1e3899","Type":"ContainerStarted","Data":"a4ba356c97155f58e445c6b2e8b6ee1d387b893ed9a4aa635938b6736c1c91d7"} Jan 21 18:00:54 crc kubenswrapper[4808]: I0121 18:00:54.384225 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 21 18:00:54 crc kubenswrapper[4808]: I0121 18:00:54.887461 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 21 18:00:55 crc kubenswrapper[4808]: I0121 18:00:55.041566 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 21 18:00:55 crc kubenswrapper[4808]: I0121 18:00:55.320410 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" event={"ID":"8e95cdda-a4de-4569-8c2e-dd9eab1e3899","Type":"ContainerStarted","Data":"41a1443c63ca8eb7e31899349b124333f428f306d39cc78126755e8cff665a37"} Jan 21 18:00:55 crc kubenswrapper[4808]: I0121 18:00:55.320743 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:55 crc kubenswrapper[4808]: I0121 18:00:55.325485 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" Jan 21 18:00:55 crc kubenswrapper[4808]: I0121 18:00:55.332645 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 21 18:00:55 crc kubenswrapper[4808]: I0121 18:00:55.340635 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-756fdfd6b5-hjcv2" podStartSLOduration=54.340616208 podStartE2EDuration="54.340616208s" podCreationTimestamp="2026-01-21 18:00:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 18:00:55.339093289 +0000 UTC m=+290.069604194" watchObservedRunningTime="2026-01-21 18:00:55.340616208 +0000 UTC m=+290.071127123" Jan 21 18:00:55 crc kubenswrapper[4808]: I0121 18:00:55.566845 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 21 18:00:56 crc kubenswrapper[4808]: I0121 18:00:56.350593 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 21 18:00:57 crc kubenswrapper[4808]: I0121 18:00:57.447643 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 21 18:01:02 crc kubenswrapper[4808]: I0121 18:01:02.268747 4808 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 21 18:01:02 crc kubenswrapper[4808]: I0121 18:01:02.269315 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://b34dffa09c0a13c935c5d49dd0f7ef7c07c1b0d9e3a687e46ca8535fb551870a" gracePeriod=5 Jan 21 18:01:05 crc kubenswrapper[4808]: I0121 18:01:05.406120 4808 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Jan 21 18:01:07 crc kubenswrapper[4808]: I0121 18:01:07.400994 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 21 18:01:07 crc kubenswrapper[4808]: I0121 18:01:07.401489 4808 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="b34dffa09c0a13c935c5d49dd0f7ef7c07c1b0d9e3a687e46ca8535fb551870a" exitCode=137 Jan 21 18:01:08 crc kubenswrapper[4808]: I0121 18:01:08.121938 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 21 18:01:08 crc kubenswrapper[4808]: I0121 18:01:08.122018 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 18:01:08 crc kubenswrapper[4808]: I0121 18:01:08.312275 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 21 18:01:08 crc kubenswrapper[4808]: I0121 18:01:08.312338 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 21 18:01:08 crc kubenswrapper[4808]: I0121 18:01:08.312400 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 21 18:01:08 crc kubenswrapper[4808]: I0121 18:01:08.312437 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:01:08 crc kubenswrapper[4808]: I0121 18:01:08.312509 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 21 18:01:08 crc kubenswrapper[4808]: I0121 18:01:08.312515 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:01:08 crc kubenswrapper[4808]: I0121 18:01:08.312538 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 21 18:01:08 crc kubenswrapper[4808]: I0121 18:01:08.312546 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:01:08 crc kubenswrapper[4808]: I0121 18:01:08.312576 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:01:08 crc kubenswrapper[4808]: I0121 18:01:08.312796 4808 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:08 crc kubenswrapper[4808]: I0121 18:01:08.312824 4808 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:08 crc kubenswrapper[4808]: I0121 18:01:08.312840 4808 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:08 crc kubenswrapper[4808]: I0121 18:01:08.312855 4808 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:08 crc kubenswrapper[4808]: I0121 18:01:08.325280 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:01:08 crc kubenswrapper[4808]: I0121 18:01:08.409108 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 21 18:01:08 crc kubenswrapper[4808]: I0121 18:01:08.409223 4808 scope.go:117] "RemoveContainer" containerID="b34dffa09c0a13c935c5d49dd0f7ef7c07c1b0d9e3a687e46ca8535fb551870a" Jan 21 18:01:08 crc kubenswrapper[4808]: I0121 18:01:08.409337 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 18:01:08 crc kubenswrapper[4808]: I0121 18:01:08.414158 4808 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:09 crc kubenswrapper[4808]: I0121 18:01:09.531734 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Jan 21 18:01:09 crc kubenswrapper[4808]: I0121 18:01:09.533521 4808 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Jan 21 18:01:09 crc kubenswrapper[4808]: I0121 18:01:09.545054 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 21 18:01:09 crc kubenswrapper[4808]: I0121 18:01:09.545384 4808 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="046987cb-2809-411a-947c-14f9da2e57f4" Jan 21 18:01:09 crc kubenswrapper[4808]: I0121 18:01:09.547949 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 21 18:01:09 crc kubenswrapper[4808]: I0121 18:01:09.547994 4808 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="046987cb-2809-411a-947c-14f9da2e57f4" Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.345094 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8h66x"] Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.345860 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" podUID="3919133e-aaba-4dc1-b4cb-6b09f6440221" containerName="controller-manager" containerID="cri-o://4c6a75571185d2caa1a47120b61b4c6781115c979189cb3b908243d3d19c82b5" gracePeriod=30 Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.386349 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8"] Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.386629 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" podUID="7af3c1d3-6e1a-4442-96ab-0aefbac80ab5" containerName="route-controller-manager" containerID="cri-o://c0cabeb627bf3a86c247f34d3e1a6343af8dcaadbc9bf25abef9963c2a42a8e4" gracePeriod=30 Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.818878 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.823859 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.936465 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3919133e-aaba-4dc1-b4cb-6b09f6440221-proxy-ca-bundles\") pod \"3919133e-aaba-4dc1-b4cb-6b09f6440221\" (UID: \"3919133e-aaba-4dc1-b4cb-6b09f6440221\") " Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.936593 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-config\") pod \"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5\" (UID: \"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5\") " Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.936653 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3919133e-aaba-4dc1-b4cb-6b09f6440221-config\") pod \"3919133e-aaba-4dc1-b4cb-6b09f6440221\" (UID: \"3919133e-aaba-4dc1-b4cb-6b09f6440221\") " Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.936681 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-serving-cert\") pod \"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5\" (UID: \"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5\") " Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.936703 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-client-ca\") pod \"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5\" (UID: \"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5\") " Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.936731 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3919133e-aaba-4dc1-b4cb-6b09f6440221-serving-cert\") pod \"3919133e-aaba-4dc1-b4cb-6b09f6440221\" (UID: \"3919133e-aaba-4dc1-b4cb-6b09f6440221\") " Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.936755 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-589nd\" (UniqueName: \"kubernetes.io/projected/3919133e-aaba-4dc1-b4cb-6b09f6440221-kube-api-access-589nd\") pod \"3919133e-aaba-4dc1-b4cb-6b09f6440221\" (UID: \"3919133e-aaba-4dc1-b4cb-6b09f6440221\") " Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.936792 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3919133e-aaba-4dc1-b4cb-6b09f6440221-client-ca\") pod \"3919133e-aaba-4dc1-b4cb-6b09f6440221\" (UID: \"3919133e-aaba-4dc1-b4cb-6b09f6440221\") " Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.936824 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zt6\" (UniqueName: \"kubernetes.io/projected/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-kube-api-access-x7zt6\") pod \"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5\" (UID: \"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5\") " Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.937579 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-client-ca" (OuterVolumeSpecName: "client-ca") pod "7af3c1d3-6e1a-4442-96ab-0aefbac80ab5" (UID: "7af3c1d3-6e1a-4442-96ab-0aefbac80ab5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.937632 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3919133e-aaba-4dc1-b4cb-6b09f6440221-config" (OuterVolumeSpecName: "config") pod "3919133e-aaba-4dc1-b4cb-6b09f6440221" (UID: "3919133e-aaba-4dc1-b4cb-6b09f6440221"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.937600 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3919133e-aaba-4dc1-b4cb-6b09f6440221-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "3919133e-aaba-4dc1-b4cb-6b09f6440221" (UID: "3919133e-aaba-4dc1-b4cb-6b09f6440221"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.937733 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-config" (OuterVolumeSpecName: "config") pod "7af3c1d3-6e1a-4442-96ab-0aefbac80ab5" (UID: "7af3c1d3-6e1a-4442-96ab-0aefbac80ab5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.937868 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3919133e-aaba-4dc1-b4cb-6b09f6440221-client-ca" (OuterVolumeSpecName: "client-ca") pod "3919133e-aaba-4dc1-b4cb-6b09f6440221" (UID: "3919133e-aaba-4dc1-b4cb-6b09f6440221"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.942603 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-kube-api-access-x7zt6" (OuterVolumeSpecName: "kube-api-access-x7zt6") pod "7af3c1d3-6e1a-4442-96ab-0aefbac80ab5" (UID: "7af3c1d3-6e1a-4442-96ab-0aefbac80ab5"). InnerVolumeSpecName "kube-api-access-x7zt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.942861 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7af3c1d3-6e1a-4442-96ab-0aefbac80ab5" (UID: "7af3c1d3-6e1a-4442-96ab-0aefbac80ab5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.942925 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3919133e-aaba-4dc1-b4cb-6b09f6440221-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3919133e-aaba-4dc1-b4cb-6b09f6440221" (UID: "3919133e-aaba-4dc1-b4cb-6b09f6440221"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:01:14 crc kubenswrapper[4808]: I0121 18:01:14.943160 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3919133e-aaba-4dc1-b4cb-6b09f6440221-kube-api-access-589nd" (OuterVolumeSpecName: "kube-api-access-589nd") pod "3919133e-aaba-4dc1-b4cb-6b09f6440221" (UID: "3919133e-aaba-4dc1-b4cb-6b09f6440221"). InnerVolumeSpecName "kube-api-access-589nd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.037727 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3919133e-aaba-4dc1-b4cb-6b09f6440221-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.037768 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-589nd\" (UniqueName: \"kubernetes.io/projected/3919133e-aaba-4dc1-b4cb-6b09f6440221-kube-api-access-589nd\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.037785 4808 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3919133e-aaba-4dc1-b4cb-6b09f6440221-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.037796 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zt6\" (UniqueName: \"kubernetes.io/projected/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-kube-api-access-x7zt6\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.037806 4808 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3919133e-aaba-4dc1-b4cb-6b09f6440221-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.037819 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-config\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.037828 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3919133e-aaba-4dc1-b4cb-6b09f6440221-config\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.037838 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.037847 4808 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.458215 4808 generic.go:334] "Generic (PLEG): container finished" podID="7af3c1d3-6e1a-4442-96ab-0aefbac80ab5" containerID="c0cabeb627bf3a86c247f34d3e1a6343af8dcaadbc9bf25abef9963c2a42a8e4" exitCode=0 Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.458312 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" event={"ID":"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5","Type":"ContainerDied","Data":"c0cabeb627bf3a86c247f34d3e1a6343af8dcaadbc9bf25abef9963c2a42a8e4"} Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.458485 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.458943 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8" event={"ID":"7af3c1d3-6e1a-4442-96ab-0aefbac80ab5","Type":"ContainerDied","Data":"7679e7a24ca34b070746c965775003ef9ae5bd5ca74a3c7385ded5a88b316e30"} Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.459003 4808 scope.go:117] "RemoveContainer" containerID="c0cabeb627bf3a86c247f34d3e1a6343af8dcaadbc9bf25abef9963c2a42a8e4" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.462882 4808 generic.go:334] "Generic (PLEG): container finished" podID="3919133e-aaba-4dc1-b4cb-6b09f6440221" containerID="4c6a75571185d2caa1a47120b61b4c6781115c979189cb3b908243d3d19c82b5" exitCode=0 Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.462924 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" event={"ID":"3919133e-aaba-4dc1-b4cb-6b09f6440221","Type":"ContainerDied","Data":"4c6a75571185d2caa1a47120b61b4c6781115c979189cb3b908243d3d19c82b5"} Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.462941 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" event={"ID":"3919133e-aaba-4dc1-b4cb-6b09f6440221","Type":"ContainerDied","Data":"0fafb665cfc5ffb6789f9e9c4f7e96b7efe3a1a1c6c27490504d9a8d24447553"} Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.463003 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8h66x" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.526212 4808 scope.go:117] "RemoveContainer" containerID="c0cabeb627bf3a86c247f34d3e1a6343af8dcaadbc9bf25abef9963c2a42a8e4" Jan 21 18:01:15 crc kubenswrapper[4808]: E0121 18:01:15.526681 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0cabeb627bf3a86c247f34d3e1a6343af8dcaadbc9bf25abef9963c2a42a8e4\": container with ID starting with c0cabeb627bf3a86c247f34d3e1a6343af8dcaadbc9bf25abef9963c2a42a8e4 not found: ID does not exist" containerID="c0cabeb627bf3a86c247f34d3e1a6343af8dcaadbc9bf25abef9963c2a42a8e4" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.526715 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0cabeb627bf3a86c247f34d3e1a6343af8dcaadbc9bf25abef9963c2a42a8e4"} err="failed to get container status \"c0cabeb627bf3a86c247f34d3e1a6343af8dcaadbc9bf25abef9963c2a42a8e4\": rpc error: code = NotFound desc = could not find container \"c0cabeb627bf3a86c247f34d3e1a6343af8dcaadbc9bf25abef9963c2a42a8e4\": container with ID starting with c0cabeb627bf3a86c247f34d3e1a6343af8dcaadbc9bf25abef9963c2a42a8e4 not found: ID does not exist" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.526740 4808 scope.go:117] "RemoveContainer" containerID="4c6a75571185d2caa1a47120b61b4c6781115c979189cb3b908243d3d19c82b5" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.546720 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8h66x"] Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.551078 4808 scope.go:117] "RemoveContainer" containerID="4c6a75571185d2caa1a47120b61b4c6781115c979189cb3b908243d3d19c82b5" Jan 21 18:01:15 crc kubenswrapper[4808]: E0121 18:01:15.551749 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c6a75571185d2caa1a47120b61b4c6781115c979189cb3b908243d3d19c82b5\": container with ID starting with 4c6a75571185d2caa1a47120b61b4c6781115c979189cb3b908243d3d19c82b5 not found: ID does not exist" containerID="4c6a75571185d2caa1a47120b61b4c6781115c979189cb3b908243d3d19c82b5" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.551812 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c6a75571185d2caa1a47120b61b4c6781115c979189cb3b908243d3d19c82b5"} err="failed to get container status \"4c6a75571185d2caa1a47120b61b4c6781115c979189cb3b908243d3d19c82b5\": rpc error: code = NotFound desc = could not find container \"4c6a75571185d2caa1a47120b61b4c6781115c979189cb3b908243d3d19c82b5\": container with ID starting with 4c6a75571185d2caa1a47120b61b4c6781115c979189cb3b908243d3d19c82b5 not found: ID does not exist" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.553872 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8h66x"] Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.557945 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8"] Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.562622 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wn2m8"] Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.787178 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6"] Jan 21 18:01:15 crc kubenswrapper[4808]: E0121 18:01:15.787500 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3919133e-aaba-4dc1-b4cb-6b09f6440221" containerName="controller-manager" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.787520 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="3919133e-aaba-4dc1-b4cb-6b09f6440221" containerName="controller-manager" Jan 21 18:01:15 crc kubenswrapper[4808]: E0121 18:01:15.787537 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.787544 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 21 18:01:15 crc kubenswrapper[4808]: E0121 18:01:15.787559 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7af3c1d3-6e1a-4442-96ab-0aefbac80ab5" containerName="route-controller-manager" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.787567 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7af3c1d3-6e1a-4442-96ab-0aefbac80ab5" containerName="route-controller-manager" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.787685 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7af3c1d3-6e1a-4442-96ab-0aefbac80ab5" containerName="route-controller-manager" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.787697 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="3919133e-aaba-4dc1-b4cb-6b09f6440221" containerName="controller-manager" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.787707 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.788100 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.793660 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.794055 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.794416 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.794570 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.794692 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.794823 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.802417 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-74996fc445-x29kx"] Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.803784 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.807579 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.810565 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.810650 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.811513 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.811875 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.811932 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.814939 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6"] Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.820019 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74996fc445-x29kx"] Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.820483 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.952757 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ab26c4b-4728-4cd2-b215-e12f00f396bc-config\") pod \"route-controller-manager-648ff87cd5-7ngm6\" (UID: \"8ab26c4b-4728-4cd2-b215-e12f00f396bc\") " pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.952884 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/961a0fb2-cce4-441d-ab82-37f78562173e-proxy-ca-bundles\") pod \"controller-manager-74996fc445-x29kx\" (UID: \"961a0fb2-cce4-441d-ab82-37f78562173e\") " pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.952932 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ab26c4b-4728-4cd2-b215-e12f00f396bc-serving-cert\") pod \"route-controller-manager-648ff87cd5-7ngm6\" (UID: \"8ab26c4b-4728-4cd2-b215-e12f00f396bc\") " pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.952953 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/961a0fb2-cce4-441d-ab82-37f78562173e-client-ca\") pod \"controller-manager-74996fc445-x29kx\" (UID: \"961a0fb2-cce4-441d-ab82-37f78562173e\") " pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.953186 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/961a0fb2-cce4-441d-ab82-37f78562173e-serving-cert\") pod \"controller-manager-74996fc445-x29kx\" (UID: \"961a0fb2-cce4-441d-ab82-37f78562173e\") " pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.953347 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phvss\" (UniqueName: \"kubernetes.io/projected/961a0fb2-cce4-441d-ab82-37f78562173e-kube-api-access-phvss\") pod \"controller-manager-74996fc445-x29kx\" (UID: \"961a0fb2-cce4-441d-ab82-37f78562173e\") " pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.953426 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp7h8\" (UniqueName: \"kubernetes.io/projected/8ab26c4b-4728-4cd2-b215-e12f00f396bc-kube-api-access-fp7h8\") pod \"route-controller-manager-648ff87cd5-7ngm6\" (UID: \"8ab26c4b-4728-4cd2-b215-e12f00f396bc\") " pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.953472 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/961a0fb2-cce4-441d-ab82-37f78562173e-config\") pod \"controller-manager-74996fc445-x29kx\" (UID: \"961a0fb2-cce4-441d-ab82-37f78562173e\") " pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:15 crc kubenswrapper[4808]: I0121 18:01:15.953566 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ab26c4b-4728-4cd2-b215-e12f00f396bc-client-ca\") pod \"route-controller-manager-648ff87cd5-7ngm6\" (UID: \"8ab26c4b-4728-4cd2-b215-e12f00f396bc\") " pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.055350 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/961a0fb2-cce4-441d-ab82-37f78562173e-proxy-ca-bundles\") pod \"controller-manager-74996fc445-x29kx\" (UID: \"961a0fb2-cce4-441d-ab82-37f78562173e\") " pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.055444 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ab26c4b-4728-4cd2-b215-e12f00f396bc-serving-cert\") pod \"route-controller-manager-648ff87cd5-7ngm6\" (UID: \"8ab26c4b-4728-4cd2-b215-e12f00f396bc\") " pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.055465 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/961a0fb2-cce4-441d-ab82-37f78562173e-client-ca\") pod \"controller-manager-74996fc445-x29kx\" (UID: \"961a0fb2-cce4-441d-ab82-37f78562173e\") " pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.055527 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/961a0fb2-cce4-441d-ab82-37f78562173e-serving-cert\") pod \"controller-manager-74996fc445-x29kx\" (UID: \"961a0fb2-cce4-441d-ab82-37f78562173e\") " pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.055550 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phvss\" (UniqueName: \"kubernetes.io/projected/961a0fb2-cce4-441d-ab82-37f78562173e-kube-api-access-phvss\") pod \"controller-manager-74996fc445-x29kx\" (UID: \"961a0fb2-cce4-441d-ab82-37f78562173e\") " pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.055587 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp7h8\" (UniqueName: \"kubernetes.io/projected/8ab26c4b-4728-4cd2-b215-e12f00f396bc-kube-api-access-fp7h8\") pod \"route-controller-manager-648ff87cd5-7ngm6\" (UID: \"8ab26c4b-4728-4cd2-b215-e12f00f396bc\") " pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.055604 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/961a0fb2-cce4-441d-ab82-37f78562173e-config\") pod \"controller-manager-74996fc445-x29kx\" (UID: \"961a0fb2-cce4-441d-ab82-37f78562173e\") " pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.055624 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ab26c4b-4728-4cd2-b215-e12f00f396bc-client-ca\") pod \"route-controller-manager-648ff87cd5-7ngm6\" (UID: \"8ab26c4b-4728-4cd2-b215-e12f00f396bc\") " pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.055642 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ab26c4b-4728-4cd2-b215-e12f00f396bc-config\") pod \"route-controller-manager-648ff87cd5-7ngm6\" (UID: \"8ab26c4b-4728-4cd2-b215-e12f00f396bc\") " pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.057216 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ab26c4b-4728-4cd2-b215-e12f00f396bc-client-ca\") pod \"route-controller-manager-648ff87cd5-7ngm6\" (UID: \"8ab26c4b-4728-4cd2-b215-e12f00f396bc\") " pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.057370 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ab26c4b-4728-4cd2-b215-e12f00f396bc-config\") pod \"route-controller-manager-648ff87cd5-7ngm6\" (UID: \"8ab26c4b-4728-4cd2-b215-e12f00f396bc\") " pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.058261 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/961a0fb2-cce4-441d-ab82-37f78562173e-config\") pod \"controller-manager-74996fc445-x29kx\" (UID: \"961a0fb2-cce4-441d-ab82-37f78562173e\") " pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.058461 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/961a0fb2-cce4-441d-ab82-37f78562173e-proxy-ca-bundles\") pod \"controller-manager-74996fc445-x29kx\" (UID: \"961a0fb2-cce4-441d-ab82-37f78562173e\") " pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.058917 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/961a0fb2-cce4-441d-ab82-37f78562173e-client-ca\") pod \"controller-manager-74996fc445-x29kx\" (UID: \"961a0fb2-cce4-441d-ab82-37f78562173e\") " pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.064279 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/961a0fb2-cce4-441d-ab82-37f78562173e-serving-cert\") pod \"controller-manager-74996fc445-x29kx\" (UID: \"961a0fb2-cce4-441d-ab82-37f78562173e\") " pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.065893 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ab26c4b-4728-4cd2-b215-e12f00f396bc-serving-cert\") pod \"route-controller-manager-648ff87cd5-7ngm6\" (UID: \"8ab26c4b-4728-4cd2-b215-e12f00f396bc\") " pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.072033 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phvss\" (UniqueName: \"kubernetes.io/projected/961a0fb2-cce4-441d-ab82-37f78562173e-kube-api-access-phvss\") pod \"controller-manager-74996fc445-x29kx\" (UID: \"961a0fb2-cce4-441d-ab82-37f78562173e\") " pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.074755 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp7h8\" (UniqueName: \"kubernetes.io/projected/8ab26c4b-4728-4cd2-b215-e12f00f396bc-kube-api-access-fp7h8\") pod \"route-controller-manager-648ff87cd5-7ngm6\" (UID: \"8ab26c4b-4728-4cd2-b215-e12f00f396bc\") " pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.117532 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.138360 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.339844 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" podUID="2b3e892c-c86a-42cd-b257-d722cd033b5e" containerName="registry" containerID="cri-o://db91f40f6def2d0a37ace9509c0e5d78a6880ff28cc615fa2f892df12ac32746" gracePeriod=30 Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.357189 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74996fc445-x29kx"] Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.392336 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6"] Jan 21 18:01:16 crc kubenswrapper[4808]: W0121 18:01:16.398371 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ab26c4b_4728_4cd2_b215_e12f00f396bc.slice/crio-0e651e60388eb4c23b5fe19f446f9a8bebdd9668c7840a7a4aabb1276f1779c5 WatchSource:0}: Error finding container 0e651e60388eb4c23b5fe19f446f9a8bebdd9668c7840a7a4aabb1276f1779c5: Status 404 returned error can't find the container with id 0e651e60388eb4c23b5fe19f446f9a8bebdd9668c7840a7a4aabb1276f1779c5 Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.470130 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" event={"ID":"8ab26c4b-4728-4cd2-b215-e12f00f396bc","Type":"ContainerStarted","Data":"0e651e60388eb4c23b5fe19f446f9a8bebdd9668c7840a7a4aabb1276f1779c5"} Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.471837 4808 generic.go:334] "Generic (PLEG): container finished" podID="2b3e892c-c86a-42cd-b257-d722cd033b5e" containerID="db91f40f6def2d0a37ace9509c0e5d78a6880ff28cc615fa2f892df12ac32746" exitCode=0 Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.471902 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" event={"ID":"2b3e892c-c86a-42cd-b257-d722cd033b5e","Type":"ContainerDied","Data":"db91f40f6def2d0a37ace9509c0e5d78a6880ff28cc615fa2f892df12ac32746"} Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.474457 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" event={"ID":"961a0fb2-cce4-441d-ab82-37f78562173e","Type":"ContainerStarted","Data":"0bc465d23218a98c9503a35dca58a781d1278b559569230e625106764d4cbbaf"} Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.811657 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.971729 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"2b3e892c-c86a-42cd-b257-d722cd033b5e\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.971777 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b3e892c-c86a-42cd-b257-d722cd033b5e-ca-trust-extracted\") pod \"2b3e892c-c86a-42cd-b257-d722cd033b5e\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.971796 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b3e892c-c86a-42cd-b257-d722cd033b5e-registry-tls\") pod \"2b3e892c-c86a-42cd-b257-d722cd033b5e\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.971833 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b3e892c-c86a-42cd-b257-d722cd033b5e-bound-sa-token\") pod \"2b3e892c-c86a-42cd-b257-d722cd033b5e\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.971854 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b3e892c-c86a-42cd-b257-d722cd033b5e-trusted-ca\") pod \"2b3e892c-c86a-42cd-b257-d722cd033b5e\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.971877 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b3e892c-c86a-42cd-b257-d722cd033b5e-registry-certificates\") pod \"2b3e892c-c86a-42cd-b257-d722cd033b5e\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.971900 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b3e892c-c86a-42cd-b257-d722cd033b5e-installation-pull-secrets\") pod \"2b3e892c-c86a-42cd-b257-d722cd033b5e\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.971934 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wptxh\" (UniqueName: \"kubernetes.io/projected/2b3e892c-c86a-42cd-b257-d722cd033b5e-kube-api-access-wptxh\") pod \"2b3e892c-c86a-42cd-b257-d722cd033b5e\" (UID: \"2b3e892c-c86a-42cd-b257-d722cd033b5e\") " Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.973067 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b3e892c-c86a-42cd-b257-d722cd033b5e-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "2b3e892c-c86a-42cd-b257-d722cd033b5e" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.973088 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b3e892c-c86a-42cd-b257-d722cd033b5e-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "2b3e892c-c86a-42cd-b257-d722cd033b5e" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.977222 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b3e892c-c86a-42cd-b257-d722cd033b5e-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "2b3e892c-c86a-42cd-b257-d722cd033b5e" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.978420 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b3e892c-c86a-42cd-b257-d722cd033b5e-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "2b3e892c-c86a-42cd-b257-d722cd033b5e" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.979798 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b3e892c-c86a-42cd-b257-d722cd033b5e-kube-api-access-wptxh" (OuterVolumeSpecName: "kube-api-access-wptxh") pod "2b3e892c-c86a-42cd-b257-d722cd033b5e" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e"). InnerVolumeSpecName "kube-api-access-wptxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.981073 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "2b3e892c-c86a-42cd-b257-d722cd033b5e" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.991563 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b3e892c-c86a-42cd-b257-d722cd033b5e-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "2b3e892c-c86a-42cd-b257-d722cd033b5e" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:01:16 crc kubenswrapper[4808]: I0121 18:01:16.997334 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b3e892c-c86a-42cd-b257-d722cd033b5e-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "2b3e892c-c86a-42cd-b257-d722cd033b5e" (UID: "2b3e892c-c86a-42cd-b257-d722cd033b5e"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.073428 4808 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b3e892c-c86a-42cd-b257-d722cd033b5e-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.073478 4808 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b3e892c-c86a-42cd-b257-d722cd033b5e-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.073491 4808 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b3e892c-c86a-42cd-b257-d722cd033b5e-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.073504 4808 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b3e892c-c86a-42cd-b257-d722cd033b5e-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.073513 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wptxh\" (UniqueName: \"kubernetes.io/projected/2b3e892c-c86a-42cd-b257-d722cd033b5e-kube-api-access-wptxh\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.073522 4808 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b3e892c-c86a-42cd-b257-d722cd033b5e-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.073532 4808 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b3e892c-c86a-42cd-b257-d722cd033b5e-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.481025 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.481078 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-n56cl" event={"ID":"2b3e892c-c86a-42cd-b257-d722cd033b5e","Type":"ContainerDied","Data":"d2434840d85f0c1ca009b377e0f76661a76a8aef8c14082e80d506bd909910e6"} Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.481145 4808 scope.go:117] "RemoveContainer" containerID="db91f40f6def2d0a37ace9509c0e5d78a6880ff28cc615fa2f892df12ac32746" Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.482699 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" event={"ID":"961a0fb2-cce4-441d-ab82-37f78562173e","Type":"ContainerStarted","Data":"80c2a38df2887b274a574f57018ed4593139f0eecb763fd1222a3aeb0efa7e79"} Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.483288 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.484890 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" event={"ID":"8ab26c4b-4728-4cd2-b215-e12f00f396bc","Type":"ContainerStarted","Data":"d20814d937d700e2dba9a2b559a2a4cc4fad036d4c288d6bff4d89d17a7f1df0"} Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.485369 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.493162 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.494084 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.514500 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" podStartSLOduration=3.5144758659999997 podStartE2EDuration="3.514475866s" podCreationTimestamp="2026-01-21 18:01:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 18:01:17.511555682 +0000 UTC m=+312.242066567" watchObservedRunningTime="2026-01-21 18:01:17.514475866 +0000 UTC m=+312.244986751" Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.528877 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" podStartSLOduration=3.528859109 podStartE2EDuration="3.528859109s" podCreationTimestamp="2026-01-21 18:01:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 18:01:17.52771279 +0000 UTC m=+312.258223685" watchObservedRunningTime="2026-01-21 18:01:17.528859109 +0000 UTC m=+312.259369994" Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.530318 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3919133e-aaba-4dc1-b4cb-6b09f6440221" path="/var/lib/kubelet/pods/3919133e-aaba-4dc1-b4cb-6b09f6440221/volumes" Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.531023 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7af3c1d3-6e1a-4442-96ab-0aefbac80ab5" path="/var/lib/kubelet/pods/7af3c1d3-6e1a-4442-96ab-0aefbac80ab5/volumes" Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.595890 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-n56cl"] Jan 21 18:01:17 crc kubenswrapper[4808]: I0121 18:01:17.608369 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-n56cl"] Jan 21 18:01:19 crc kubenswrapper[4808]: I0121 18:01:19.533054 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b3e892c-c86a-42cd-b257-d722cd033b5e" path="/var/lib/kubelet/pods/2b3e892c-c86a-42cd-b257-d722cd033b5e/volumes" Jan 21 18:01:24 crc kubenswrapper[4808]: I0121 18:01:24.775104 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6m94d"] Jan 21 18:01:24 crc kubenswrapper[4808]: E0121 18:01:24.775838 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b3e892c-c86a-42cd-b257-d722cd033b5e" containerName="registry" Jan 21 18:01:24 crc kubenswrapper[4808]: I0121 18:01:24.775859 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b3e892c-c86a-42cd-b257-d722cd033b5e" containerName="registry" Jan 21 18:01:24 crc kubenswrapper[4808]: I0121 18:01:24.776033 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b3e892c-c86a-42cd-b257-d722cd033b5e" containerName="registry" Jan 21 18:01:24 crc kubenswrapper[4808]: I0121 18:01:24.777239 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6m94d" Jan 21 18:01:24 crc kubenswrapper[4808]: I0121 18:01:24.779802 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 21 18:01:24 crc kubenswrapper[4808]: I0121 18:01:24.793770 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6m94d"] Jan 21 18:01:24 crc kubenswrapper[4808]: I0121 18:01:24.835706 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62cdb877-c15b-4990-8820-3dcc6ffc0b04-catalog-content\") pod \"certified-operators-6m94d\" (UID: \"62cdb877-c15b-4990-8820-3dcc6ffc0b04\") " pod="openshift-marketplace/certified-operators-6m94d" Jan 21 18:01:24 crc kubenswrapper[4808]: I0121 18:01:24.835757 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8snm\" (UniqueName: \"kubernetes.io/projected/62cdb877-c15b-4990-8820-3dcc6ffc0b04-kube-api-access-v8snm\") pod \"certified-operators-6m94d\" (UID: \"62cdb877-c15b-4990-8820-3dcc6ffc0b04\") " pod="openshift-marketplace/certified-operators-6m94d" Jan 21 18:01:24 crc kubenswrapper[4808]: I0121 18:01:24.835840 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62cdb877-c15b-4990-8820-3dcc6ffc0b04-utilities\") pod \"certified-operators-6m94d\" (UID: \"62cdb877-c15b-4990-8820-3dcc6ffc0b04\") " pod="openshift-marketplace/certified-operators-6m94d" Jan 21 18:01:24 crc kubenswrapper[4808]: I0121 18:01:24.936618 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62cdb877-c15b-4990-8820-3dcc6ffc0b04-catalog-content\") pod \"certified-operators-6m94d\" (UID: \"62cdb877-c15b-4990-8820-3dcc6ffc0b04\") " pod="openshift-marketplace/certified-operators-6m94d" Jan 21 18:01:24 crc kubenswrapper[4808]: I0121 18:01:24.936676 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8snm\" (UniqueName: \"kubernetes.io/projected/62cdb877-c15b-4990-8820-3dcc6ffc0b04-kube-api-access-v8snm\") pod \"certified-operators-6m94d\" (UID: \"62cdb877-c15b-4990-8820-3dcc6ffc0b04\") " pod="openshift-marketplace/certified-operators-6m94d" Jan 21 18:01:24 crc kubenswrapper[4808]: I0121 18:01:24.936720 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62cdb877-c15b-4990-8820-3dcc6ffc0b04-utilities\") pod \"certified-operators-6m94d\" (UID: \"62cdb877-c15b-4990-8820-3dcc6ffc0b04\") " pod="openshift-marketplace/certified-operators-6m94d" Jan 21 18:01:24 crc kubenswrapper[4808]: I0121 18:01:24.937182 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62cdb877-c15b-4990-8820-3dcc6ffc0b04-utilities\") pod \"certified-operators-6m94d\" (UID: \"62cdb877-c15b-4990-8820-3dcc6ffc0b04\") " pod="openshift-marketplace/certified-operators-6m94d" Jan 21 18:01:24 crc kubenswrapper[4808]: I0121 18:01:24.937466 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62cdb877-c15b-4990-8820-3dcc6ffc0b04-catalog-content\") pod \"certified-operators-6m94d\" (UID: \"62cdb877-c15b-4990-8820-3dcc6ffc0b04\") " pod="openshift-marketplace/certified-operators-6m94d" Jan 21 18:01:24 crc kubenswrapper[4808]: I0121 18:01:24.966126 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8snm\" (UniqueName: \"kubernetes.io/projected/62cdb877-c15b-4990-8820-3dcc6ffc0b04-kube-api-access-v8snm\") pod \"certified-operators-6m94d\" (UID: \"62cdb877-c15b-4990-8820-3dcc6ffc0b04\") " pod="openshift-marketplace/certified-operators-6m94d" Jan 21 18:01:25 crc kubenswrapper[4808]: I0121 18:01:25.099017 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6m94d" Jan 21 18:01:25 crc kubenswrapper[4808]: I0121 18:01:25.555206 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6m94d"] Jan 21 18:01:26 crc kubenswrapper[4808]: I0121 18:01:26.559281 4808 generic.go:334] "Generic (PLEG): container finished" podID="62cdb877-c15b-4990-8820-3dcc6ffc0b04" containerID="a49b3340884343001efa7248d836b1be90456f2a185cc7ddc826cc0ce8b632b3" exitCode=0 Jan 21 18:01:26 crc kubenswrapper[4808]: I0121 18:01:26.559379 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6m94d" event={"ID":"62cdb877-c15b-4990-8820-3dcc6ffc0b04","Type":"ContainerDied","Data":"a49b3340884343001efa7248d836b1be90456f2a185cc7ddc826cc0ce8b632b3"} Jan 21 18:01:26 crc kubenswrapper[4808]: I0121 18:01:26.559625 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6m94d" event={"ID":"62cdb877-c15b-4990-8820-3dcc6ffc0b04","Type":"ContainerStarted","Data":"b5c91b870a29cb0d04d1dc9edd6b03fe5ebc7f5f85546d433b745bc38c488d4a"} Jan 21 18:01:27 crc kubenswrapper[4808]: I0121 18:01:27.159554 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-54shl"] Jan 21 18:01:27 crc kubenswrapper[4808]: I0121 18:01:27.160924 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-54shl" Jan 21 18:01:27 crc kubenswrapper[4808]: I0121 18:01:27.163522 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 21 18:01:27 crc kubenswrapper[4808]: I0121 18:01:27.169872 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-54shl"] Jan 21 18:01:27 crc kubenswrapper[4808]: I0121 18:01:27.270233 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c4c6107-02cd-47e8-b29a-88190d15f030-utilities\") pod \"redhat-operators-54shl\" (UID: \"9c4c6107-02cd-47e8-b29a-88190d15f030\") " pod="openshift-marketplace/redhat-operators-54shl" Jan 21 18:01:27 crc kubenswrapper[4808]: I0121 18:01:27.270333 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jvd5\" (UniqueName: \"kubernetes.io/projected/9c4c6107-02cd-47e8-b29a-88190d15f030-kube-api-access-7jvd5\") pod \"redhat-operators-54shl\" (UID: \"9c4c6107-02cd-47e8-b29a-88190d15f030\") " pod="openshift-marketplace/redhat-operators-54shl" Jan 21 18:01:27 crc kubenswrapper[4808]: I0121 18:01:27.270386 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c4c6107-02cd-47e8-b29a-88190d15f030-catalog-content\") pod \"redhat-operators-54shl\" (UID: \"9c4c6107-02cd-47e8-b29a-88190d15f030\") " pod="openshift-marketplace/redhat-operators-54shl" Jan 21 18:01:27 crc kubenswrapper[4808]: I0121 18:01:27.371540 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c4c6107-02cd-47e8-b29a-88190d15f030-catalog-content\") pod \"redhat-operators-54shl\" (UID: \"9c4c6107-02cd-47e8-b29a-88190d15f030\") " pod="openshift-marketplace/redhat-operators-54shl" Jan 21 18:01:27 crc kubenswrapper[4808]: I0121 18:01:27.371645 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c4c6107-02cd-47e8-b29a-88190d15f030-utilities\") pod \"redhat-operators-54shl\" (UID: \"9c4c6107-02cd-47e8-b29a-88190d15f030\") " pod="openshift-marketplace/redhat-operators-54shl" Jan 21 18:01:27 crc kubenswrapper[4808]: I0121 18:01:27.371673 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jvd5\" (UniqueName: \"kubernetes.io/projected/9c4c6107-02cd-47e8-b29a-88190d15f030-kube-api-access-7jvd5\") pod \"redhat-operators-54shl\" (UID: \"9c4c6107-02cd-47e8-b29a-88190d15f030\") " pod="openshift-marketplace/redhat-operators-54shl" Jan 21 18:01:27 crc kubenswrapper[4808]: I0121 18:01:27.372108 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c4c6107-02cd-47e8-b29a-88190d15f030-catalog-content\") pod \"redhat-operators-54shl\" (UID: \"9c4c6107-02cd-47e8-b29a-88190d15f030\") " pod="openshift-marketplace/redhat-operators-54shl" Jan 21 18:01:27 crc kubenswrapper[4808]: I0121 18:01:27.372155 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c4c6107-02cd-47e8-b29a-88190d15f030-utilities\") pod \"redhat-operators-54shl\" (UID: \"9c4c6107-02cd-47e8-b29a-88190d15f030\") " pod="openshift-marketplace/redhat-operators-54shl" Jan 21 18:01:27 crc kubenswrapper[4808]: I0121 18:01:27.391035 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jvd5\" (UniqueName: \"kubernetes.io/projected/9c4c6107-02cd-47e8-b29a-88190d15f030-kube-api-access-7jvd5\") pod \"redhat-operators-54shl\" (UID: \"9c4c6107-02cd-47e8-b29a-88190d15f030\") " pod="openshift-marketplace/redhat-operators-54shl" Jan 21 18:01:27 crc kubenswrapper[4808]: I0121 18:01:27.480092 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-54shl" Jan 21 18:01:27 crc kubenswrapper[4808]: I0121 18:01:27.586323 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6m94d" event={"ID":"62cdb877-c15b-4990-8820-3dcc6ffc0b04","Type":"ContainerStarted","Data":"d43fb31c253ba16440f65cf1b28595d677bbe5ce579b6a51afc16f3ec4218d04"} Jan 21 18:01:27 crc kubenswrapper[4808]: I0121 18:01:27.884938 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-54shl"] Jan 21 18:01:28 crc kubenswrapper[4808]: I0121 18:01:28.600332 4808 generic.go:334] "Generic (PLEG): container finished" podID="62cdb877-c15b-4990-8820-3dcc6ffc0b04" containerID="d43fb31c253ba16440f65cf1b28595d677bbe5ce579b6a51afc16f3ec4218d04" exitCode=0 Jan 21 18:01:28 crc kubenswrapper[4808]: I0121 18:01:28.600749 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6m94d" event={"ID":"62cdb877-c15b-4990-8820-3dcc6ffc0b04","Type":"ContainerDied","Data":"d43fb31c253ba16440f65cf1b28595d677bbe5ce579b6a51afc16f3ec4218d04"} Jan 21 18:01:28 crc kubenswrapper[4808]: I0121 18:01:28.603494 4808 generic.go:334] "Generic (PLEG): container finished" podID="9c4c6107-02cd-47e8-b29a-88190d15f030" containerID="6888c526abb731c40f1cf15d10582c309889ec8f8223451e65b732abcec44f50" exitCode=0 Jan 21 18:01:28 crc kubenswrapper[4808]: I0121 18:01:28.603535 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-54shl" event={"ID":"9c4c6107-02cd-47e8-b29a-88190d15f030","Type":"ContainerDied","Data":"6888c526abb731c40f1cf15d10582c309889ec8f8223451e65b732abcec44f50"} Jan 21 18:01:28 crc kubenswrapper[4808]: I0121 18:01:28.603561 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-54shl" event={"ID":"9c4c6107-02cd-47e8-b29a-88190d15f030","Type":"ContainerStarted","Data":"773fe47b988692e2e5f700bc48e0fdbd2bb7ee7add84d9c0c929136ed06f199d"} Jan 21 18:01:28 crc kubenswrapper[4808]: I0121 18:01:28.963471 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hzsq2"] Jan 21 18:01:28 crc kubenswrapper[4808]: I0121 18:01:28.965870 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hzsq2" Jan 21 18:01:28 crc kubenswrapper[4808]: I0121 18:01:28.973190 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hzsq2"] Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.098291 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f554b03-f857-4309-ba4c-91e8c607a14d-utilities\") pod \"certified-operators-hzsq2\" (UID: \"7f554b03-f857-4309-ba4c-91e8c607a14d\") " pod="openshift-marketplace/certified-operators-hzsq2" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.098488 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n7qm\" (UniqueName: \"kubernetes.io/projected/7f554b03-f857-4309-ba4c-91e8c607a14d-kube-api-access-5n7qm\") pod \"certified-operators-hzsq2\" (UID: \"7f554b03-f857-4309-ba4c-91e8c607a14d\") " pod="openshift-marketplace/certified-operators-hzsq2" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.098565 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f554b03-f857-4309-ba4c-91e8c607a14d-catalog-content\") pod \"certified-operators-hzsq2\" (UID: \"7f554b03-f857-4309-ba4c-91e8c607a14d\") " pod="openshift-marketplace/certified-operators-hzsq2" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.200299 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n7qm\" (UniqueName: \"kubernetes.io/projected/7f554b03-f857-4309-ba4c-91e8c607a14d-kube-api-access-5n7qm\") pod \"certified-operators-hzsq2\" (UID: \"7f554b03-f857-4309-ba4c-91e8c607a14d\") " pod="openshift-marketplace/certified-operators-hzsq2" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.200808 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f554b03-f857-4309-ba4c-91e8c607a14d-catalog-content\") pod \"certified-operators-hzsq2\" (UID: \"7f554b03-f857-4309-ba4c-91e8c607a14d\") " pod="openshift-marketplace/certified-operators-hzsq2" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.200917 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f554b03-f857-4309-ba4c-91e8c607a14d-utilities\") pod \"certified-operators-hzsq2\" (UID: \"7f554b03-f857-4309-ba4c-91e8c607a14d\") " pod="openshift-marketplace/certified-operators-hzsq2" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.201306 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f554b03-f857-4309-ba4c-91e8c607a14d-catalog-content\") pod \"certified-operators-hzsq2\" (UID: \"7f554b03-f857-4309-ba4c-91e8c607a14d\") " pod="openshift-marketplace/certified-operators-hzsq2" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.201428 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f554b03-f857-4309-ba4c-91e8c607a14d-utilities\") pod \"certified-operators-hzsq2\" (UID: \"7f554b03-f857-4309-ba4c-91e8c607a14d\") " pod="openshift-marketplace/certified-operators-hzsq2" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.224018 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n7qm\" (UniqueName: \"kubernetes.io/projected/7f554b03-f857-4309-ba4c-91e8c607a14d-kube-api-access-5n7qm\") pod \"certified-operators-hzsq2\" (UID: \"7f554b03-f857-4309-ba4c-91e8c607a14d\") " pod="openshift-marketplace/certified-operators-hzsq2" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.333832 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hzsq2" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.572687 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-59dvj"] Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.574031 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-59dvj" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.577575 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-59dvj"] Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.608981 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6m94d" event={"ID":"62cdb877-c15b-4990-8820-3dcc6ffc0b04","Type":"ContainerStarted","Data":"adf91b787ad2ac75272151d18611c4d9cbaf09e66530001a605d177fa709ff7c"} Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.610951 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-54shl" event={"ID":"9c4c6107-02cd-47e8-b29a-88190d15f030","Type":"ContainerStarted","Data":"f6f7e07e58f46da9e86457fe0d58e3b5a1677ac6848ce15535a5b70fbfb11ee3"} Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.641785 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6m94d" podStartSLOduration=2.92410001 podStartE2EDuration="5.641703014s" podCreationTimestamp="2026-01-21 18:01:24 +0000 UTC" firstStartedPulling="2026-01-21 18:01:26.5641639 +0000 UTC m=+321.294674825" lastFinishedPulling="2026-01-21 18:01:29.281766934 +0000 UTC m=+324.012277829" observedRunningTime="2026-01-21 18:01:29.63164297 +0000 UTC m=+324.362153875" watchObservedRunningTime="2026-01-21 18:01:29.641703014 +0000 UTC m=+324.372213909" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.707466 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgl66\" (UniqueName: \"kubernetes.io/projected/bc01e22f-4eb3-46d9-8bba-faffce8f294e-kube-api-access-zgl66\") pod \"redhat-operators-59dvj\" (UID: \"bc01e22f-4eb3-46d9-8bba-faffce8f294e\") " pod="openshift-marketplace/redhat-operators-59dvj" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.707541 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc01e22f-4eb3-46d9-8bba-faffce8f294e-utilities\") pod \"redhat-operators-59dvj\" (UID: \"bc01e22f-4eb3-46d9-8bba-faffce8f294e\") " pod="openshift-marketplace/redhat-operators-59dvj" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.707724 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc01e22f-4eb3-46d9-8bba-faffce8f294e-catalog-content\") pod \"redhat-operators-59dvj\" (UID: \"bc01e22f-4eb3-46d9-8bba-faffce8f294e\") " pod="openshift-marketplace/redhat-operators-59dvj" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.748143 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hzsq2"] Jan 21 18:01:29 crc kubenswrapper[4808]: W0121 18:01:29.751987 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f554b03_f857_4309_ba4c_91e8c607a14d.slice/crio-31a325e79109c191d7fd701061369dcb0bf66e6bc033bf9b6b7225d61cdb4bd6 WatchSource:0}: Error finding container 31a325e79109c191d7fd701061369dcb0bf66e6bc033bf9b6b7225d61cdb4bd6: Status 404 returned error can't find the container with id 31a325e79109c191d7fd701061369dcb0bf66e6bc033bf9b6b7225d61cdb4bd6 Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.809091 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgl66\" (UniqueName: \"kubernetes.io/projected/bc01e22f-4eb3-46d9-8bba-faffce8f294e-kube-api-access-zgl66\") pod \"redhat-operators-59dvj\" (UID: \"bc01e22f-4eb3-46d9-8bba-faffce8f294e\") " pod="openshift-marketplace/redhat-operators-59dvj" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.809173 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc01e22f-4eb3-46d9-8bba-faffce8f294e-utilities\") pod \"redhat-operators-59dvj\" (UID: \"bc01e22f-4eb3-46d9-8bba-faffce8f294e\") " pod="openshift-marketplace/redhat-operators-59dvj" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.809273 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc01e22f-4eb3-46d9-8bba-faffce8f294e-catalog-content\") pod \"redhat-operators-59dvj\" (UID: \"bc01e22f-4eb3-46d9-8bba-faffce8f294e\") " pod="openshift-marketplace/redhat-operators-59dvj" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.809645 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc01e22f-4eb3-46d9-8bba-faffce8f294e-utilities\") pod \"redhat-operators-59dvj\" (UID: \"bc01e22f-4eb3-46d9-8bba-faffce8f294e\") " pod="openshift-marketplace/redhat-operators-59dvj" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.809696 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc01e22f-4eb3-46d9-8bba-faffce8f294e-catalog-content\") pod \"redhat-operators-59dvj\" (UID: \"bc01e22f-4eb3-46d9-8bba-faffce8f294e\") " pod="openshift-marketplace/redhat-operators-59dvj" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.828152 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgl66\" (UniqueName: \"kubernetes.io/projected/bc01e22f-4eb3-46d9-8bba-faffce8f294e-kube-api-access-zgl66\") pod \"redhat-operators-59dvj\" (UID: \"bc01e22f-4eb3-46d9-8bba-faffce8f294e\") " pod="openshift-marketplace/redhat-operators-59dvj" Jan 21 18:01:29 crc kubenswrapper[4808]: I0121 18:01:29.892561 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-59dvj" Jan 21 18:01:30 crc kubenswrapper[4808]: I0121 18:01:30.309052 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-59dvj"] Jan 21 18:01:30 crc kubenswrapper[4808]: W0121 18:01:30.318123 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc01e22f_4eb3_46d9_8bba_faffce8f294e.slice/crio-c5919407149198cecf2cf9678cf162fe3e88181413cf97ab05711c3b88da63be WatchSource:0}: Error finding container c5919407149198cecf2cf9678cf162fe3e88181413cf97ab05711c3b88da63be: Status 404 returned error can't find the container with id c5919407149198cecf2cf9678cf162fe3e88181413cf97ab05711c3b88da63be Jan 21 18:01:30 crc kubenswrapper[4808]: I0121 18:01:30.617892 4808 generic.go:334] "Generic (PLEG): container finished" podID="9c4c6107-02cd-47e8-b29a-88190d15f030" containerID="f6f7e07e58f46da9e86457fe0d58e3b5a1677ac6848ce15535a5b70fbfb11ee3" exitCode=0 Jan 21 18:01:30 crc kubenswrapper[4808]: I0121 18:01:30.617954 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-54shl" event={"ID":"9c4c6107-02cd-47e8-b29a-88190d15f030","Type":"ContainerDied","Data":"f6f7e07e58f46da9e86457fe0d58e3b5a1677ac6848ce15535a5b70fbfb11ee3"} Jan 21 18:01:30 crc kubenswrapper[4808]: I0121 18:01:30.629346 4808 generic.go:334] "Generic (PLEG): container finished" podID="bc01e22f-4eb3-46d9-8bba-faffce8f294e" containerID="4c28b17f998814acb91c6d05fc94a0d4404ff2e72dec2302ff7e31367534d37d" exitCode=0 Jan 21 18:01:30 crc kubenswrapper[4808]: I0121 18:01:30.629644 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59dvj" event={"ID":"bc01e22f-4eb3-46d9-8bba-faffce8f294e","Type":"ContainerDied","Data":"4c28b17f998814acb91c6d05fc94a0d4404ff2e72dec2302ff7e31367534d37d"} Jan 21 18:01:30 crc kubenswrapper[4808]: I0121 18:01:30.629727 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59dvj" event={"ID":"bc01e22f-4eb3-46d9-8bba-faffce8f294e","Type":"ContainerStarted","Data":"c5919407149198cecf2cf9678cf162fe3e88181413cf97ab05711c3b88da63be"} Jan 21 18:01:30 crc kubenswrapper[4808]: I0121 18:01:30.634748 4808 generic.go:334] "Generic (PLEG): container finished" podID="7f554b03-f857-4309-ba4c-91e8c607a14d" containerID="1c2540cfd9f6f2aa5e9c8226454898bf6f7394cd1609b32776dbc9ddc106fc74" exitCode=0 Jan 21 18:01:30 crc kubenswrapper[4808]: I0121 18:01:30.635497 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hzsq2" event={"ID":"7f554b03-f857-4309-ba4c-91e8c607a14d","Type":"ContainerDied","Data":"1c2540cfd9f6f2aa5e9c8226454898bf6f7394cd1609b32776dbc9ddc106fc74"} Jan 21 18:01:30 crc kubenswrapper[4808]: I0121 18:01:30.635635 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hzsq2" event={"ID":"7f554b03-f857-4309-ba4c-91e8c607a14d","Type":"ContainerStarted","Data":"31a325e79109c191d7fd701061369dcb0bf66e6bc033bf9b6b7225d61cdb4bd6"} Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.376443 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xstv7"] Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.381450 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xstv7" Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.383719 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xstv7"] Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.543847 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds56w\" (UniqueName: \"kubernetes.io/projected/173b6b54-1024-4806-8056-985010126525-kube-api-access-ds56w\") pod \"certified-operators-xstv7\" (UID: \"173b6b54-1024-4806-8056-985010126525\") " pod="openshift-marketplace/certified-operators-xstv7" Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.543906 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/173b6b54-1024-4806-8056-985010126525-catalog-content\") pod \"certified-operators-xstv7\" (UID: \"173b6b54-1024-4806-8056-985010126525\") " pod="openshift-marketplace/certified-operators-xstv7" Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.544206 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/173b6b54-1024-4806-8056-985010126525-utilities\") pod \"certified-operators-xstv7\" (UID: \"173b6b54-1024-4806-8056-985010126525\") " pod="openshift-marketplace/certified-operators-xstv7" Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.641798 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59dvj" event={"ID":"bc01e22f-4eb3-46d9-8bba-faffce8f294e","Type":"ContainerStarted","Data":"3ed9c4bb797ebc91963e2e7239f9944906a3072e1b7595cc2061bed3433e7f69"} Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.643600 4808 generic.go:334] "Generic (PLEG): container finished" podID="7f554b03-f857-4309-ba4c-91e8c607a14d" containerID="6624bc1e34644ec644da76e9ba8e3fedf50bd4481db75a8fc10f7b987cf03d4a" exitCode=0 Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.643650 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hzsq2" event={"ID":"7f554b03-f857-4309-ba4c-91e8c607a14d","Type":"ContainerDied","Data":"6624bc1e34644ec644da76e9ba8e3fedf50bd4481db75a8fc10f7b987cf03d4a"} Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.645320 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/173b6b54-1024-4806-8056-985010126525-utilities\") pod \"certified-operators-xstv7\" (UID: \"173b6b54-1024-4806-8056-985010126525\") " pod="openshift-marketplace/certified-operators-xstv7" Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.645390 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds56w\" (UniqueName: \"kubernetes.io/projected/173b6b54-1024-4806-8056-985010126525-kube-api-access-ds56w\") pod \"certified-operators-xstv7\" (UID: \"173b6b54-1024-4806-8056-985010126525\") " pod="openshift-marketplace/certified-operators-xstv7" Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.645417 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/173b6b54-1024-4806-8056-985010126525-catalog-content\") pod \"certified-operators-xstv7\" (UID: \"173b6b54-1024-4806-8056-985010126525\") " pod="openshift-marketplace/certified-operators-xstv7" Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.645865 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/173b6b54-1024-4806-8056-985010126525-catalog-content\") pod \"certified-operators-xstv7\" (UID: \"173b6b54-1024-4806-8056-985010126525\") " pod="openshift-marketplace/certified-operators-xstv7" Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.646163 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/173b6b54-1024-4806-8056-985010126525-utilities\") pod \"certified-operators-xstv7\" (UID: \"173b6b54-1024-4806-8056-985010126525\") " pod="openshift-marketplace/certified-operators-xstv7" Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.647219 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-54shl" event={"ID":"9c4c6107-02cd-47e8-b29a-88190d15f030","Type":"ContainerStarted","Data":"3b5e007d295d7929bd34ff5bf413db6b05e1b68633e9312adaf53f3e570762c9"} Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.669360 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds56w\" (UniqueName: \"kubernetes.io/projected/173b6b54-1024-4806-8056-985010126525-kube-api-access-ds56w\") pod \"certified-operators-xstv7\" (UID: \"173b6b54-1024-4806-8056-985010126525\") " pod="openshift-marketplace/certified-operators-xstv7" Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.700122 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xstv7" Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.706500 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-54shl" podStartSLOduration=2.231428574 podStartE2EDuration="4.70647708s" podCreationTimestamp="2026-01-21 18:01:27 +0000 UTC" firstStartedPulling="2026-01-21 18:01:28.6045263 +0000 UTC m=+323.335037185" lastFinishedPulling="2026-01-21 18:01:31.079574806 +0000 UTC m=+325.810085691" observedRunningTime="2026-01-21 18:01:31.706204293 +0000 UTC m=+326.436715188" watchObservedRunningTime="2026-01-21 18:01:31.70647708 +0000 UTC m=+326.436987965" Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.964964 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tj8vd"] Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.966713 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tj8vd" Jan 21 18:01:31 crc kubenswrapper[4808]: I0121 18:01:31.976301 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tj8vd"] Jan 21 18:01:32 crc kubenswrapper[4808]: I0121 18:01:32.120752 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xstv7"] Jan 21 18:01:32 crc kubenswrapper[4808]: I0121 18:01:32.152543 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc0321c-4482-40a4-92e3-67f5f78f5c6e-catalog-content\") pod \"redhat-operators-tj8vd\" (UID: \"afc0321c-4482-40a4-92e3-67f5f78f5c6e\") " pod="openshift-marketplace/redhat-operators-tj8vd" Jan 21 18:01:32 crc kubenswrapper[4808]: I0121 18:01:32.152627 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgrxn\" (UniqueName: \"kubernetes.io/projected/afc0321c-4482-40a4-92e3-67f5f78f5c6e-kube-api-access-jgrxn\") pod \"redhat-operators-tj8vd\" (UID: \"afc0321c-4482-40a4-92e3-67f5f78f5c6e\") " pod="openshift-marketplace/redhat-operators-tj8vd" Jan 21 18:01:32 crc kubenswrapper[4808]: I0121 18:01:32.153189 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc0321c-4482-40a4-92e3-67f5f78f5c6e-utilities\") pod \"redhat-operators-tj8vd\" (UID: \"afc0321c-4482-40a4-92e3-67f5f78f5c6e\") " pod="openshift-marketplace/redhat-operators-tj8vd" Jan 21 18:01:32 crc kubenswrapper[4808]: I0121 18:01:32.254347 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc0321c-4482-40a4-92e3-67f5f78f5c6e-utilities\") pod \"redhat-operators-tj8vd\" (UID: \"afc0321c-4482-40a4-92e3-67f5f78f5c6e\") " pod="openshift-marketplace/redhat-operators-tj8vd" Jan 21 18:01:32 crc kubenswrapper[4808]: I0121 18:01:32.254407 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc0321c-4482-40a4-92e3-67f5f78f5c6e-catalog-content\") pod \"redhat-operators-tj8vd\" (UID: \"afc0321c-4482-40a4-92e3-67f5f78f5c6e\") " pod="openshift-marketplace/redhat-operators-tj8vd" Jan 21 18:01:32 crc kubenswrapper[4808]: I0121 18:01:32.254441 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgrxn\" (UniqueName: \"kubernetes.io/projected/afc0321c-4482-40a4-92e3-67f5f78f5c6e-kube-api-access-jgrxn\") pod \"redhat-operators-tj8vd\" (UID: \"afc0321c-4482-40a4-92e3-67f5f78f5c6e\") " pod="openshift-marketplace/redhat-operators-tj8vd" Jan 21 18:01:32 crc kubenswrapper[4808]: I0121 18:01:32.254991 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc0321c-4482-40a4-92e3-67f5f78f5c6e-utilities\") pod \"redhat-operators-tj8vd\" (UID: \"afc0321c-4482-40a4-92e3-67f5f78f5c6e\") " pod="openshift-marketplace/redhat-operators-tj8vd" Jan 21 18:01:32 crc kubenswrapper[4808]: I0121 18:01:32.255449 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc0321c-4482-40a4-92e3-67f5f78f5c6e-catalog-content\") pod \"redhat-operators-tj8vd\" (UID: \"afc0321c-4482-40a4-92e3-67f5f78f5c6e\") " pod="openshift-marketplace/redhat-operators-tj8vd" Jan 21 18:01:32 crc kubenswrapper[4808]: I0121 18:01:32.276096 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgrxn\" (UniqueName: \"kubernetes.io/projected/afc0321c-4482-40a4-92e3-67f5f78f5c6e-kube-api-access-jgrxn\") pod \"redhat-operators-tj8vd\" (UID: \"afc0321c-4482-40a4-92e3-67f5f78f5c6e\") " pod="openshift-marketplace/redhat-operators-tj8vd" Jan 21 18:01:32 crc kubenswrapper[4808]: I0121 18:01:32.293691 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tj8vd" Jan 21 18:01:32 crc kubenswrapper[4808]: I0121 18:01:32.654572 4808 generic.go:334] "Generic (PLEG): container finished" podID="bc01e22f-4eb3-46d9-8bba-faffce8f294e" containerID="3ed9c4bb797ebc91963e2e7239f9944906a3072e1b7595cc2061bed3433e7f69" exitCode=0 Jan 21 18:01:32 crc kubenswrapper[4808]: I0121 18:01:32.654621 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59dvj" event={"ID":"bc01e22f-4eb3-46d9-8bba-faffce8f294e","Type":"ContainerDied","Data":"3ed9c4bb797ebc91963e2e7239f9944906a3072e1b7595cc2061bed3433e7f69"} Jan 21 18:01:32 crc kubenswrapper[4808]: I0121 18:01:32.657218 4808 generic.go:334] "Generic (PLEG): container finished" podID="173b6b54-1024-4806-8056-985010126525" containerID="1718c9d40462fd7d04dc435d2d6e3b79508ec08e1baad202f6eae85e056b2708" exitCode=0 Jan 21 18:01:32 crc kubenswrapper[4808]: I0121 18:01:32.657292 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xstv7" event={"ID":"173b6b54-1024-4806-8056-985010126525","Type":"ContainerDied","Data":"1718c9d40462fd7d04dc435d2d6e3b79508ec08e1baad202f6eae85e056b2708"} Jan 21 18:01:32 crc kubenswrapper[4808]: I0121 18:01:32.657319 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xstv7" event={"ID":"173b6b54-1024-4806-8056-985010126525","Type":"ContainerStarted","Data":"2b789d2ed58e2ed957ab68a4d09ed6f179d74417c4b7179e1a9c082eb2e0e08c"} Jan 21 18:01:32 crc kubenswrapper[4808]: I0121 18:01:32.661468 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hzsq2" event={"ID":"7f554b03-f857-4309-ba4c-91e8c607a14d","Type":"ContainerStarted","Data":"ba1dbe7c160275afae696b07a62f00b7eb61a955782adf02caefbd95939b1493"} Jan 21 18:01:32 crc kubenswrapper[4808]: I0121 18:01:32.697532 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hzsq2" podStartSLOduration=3.06582714 podStartE2EDuration="4.69750675s" podCreationTimestamp="2026-01-21 18:01:28 +0000 UTC" firstStartedPulling="2026-01-21 18:01:30.636396216 +0000 UTC m=+325.366907101" lastFinishedPulling="2026-01-21 18:01:32.268075826 +0000 UTC m=+326.998586711" observedRunningTime="2026-01-21 18:01:32.695184803 +0000 UTC m=+327.425695698" watchObservedRunningTime="2026-01-21 18:01:32.69750675 +0000 UTC m=+327.428017635" Jan 21 18:01:32 crc kubenswrapper[4808]: I0121 18:01:32.767387 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tj8vd"] Jan 21 18:01:32 crc kubenswrapper[4808]: W0121 18:01:32.773317 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafc0321c_4482_40a4_92e3_67f5f78f5c6e.slice/crio-f0ea2fe95e73bfa4c9e3423e7f55e70ac2220bd9458d8fc65fb38105aa2d7059 WatchSource:0}: Error finding container f0ea2fe95e73bfa4c9e3423e7f55e70ac2220bd9458d8fc65fb38105aa2d7059: Status 404 returned error can't find the container with id f0ea2fe95e73bfa4c9e3423e7f55e70ac2220bd9458d8fc65fb38105aa2d7059 Jan 21 18:01:33 crc kubenswrapper[4808]: I0121 18:01:33.669523 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xstv7" event={"ID":"173b6b54-1024-4806-8056-985010126525","Type":"ContainerStarted","Data":"5202c43b50f1265bbd8a3ed0aa836614a8f0ddc4a82081f921e026f26dd76b02"} Jan 21 18:01:33 crc kubenswrapper[4808]: I0121 18:01:33.671602 4808 generic.go:334] "Generic (PLEG): container finished" podID="afc0321c-4482-40a4-92e3-67f5f78f5c6e" containerID="a98a31a17ef56ed0fdf14b9ded7d0ee27efac0f5c477ba1f52758666224737e6" exitCode=0 Jan 21 18:01:33 crc kubenswrapper[4808]: I0121 18:01:33.671680 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tj8vd" event={"ID":"afc0321c-4482-40a4-92e3-67f5f78f5c6e","Type":"ContainerDied","Data":"a98a31a17ef56ed0fdf14b9ded7d0ee27efac0f5c477ba1f52758666224737e6"} Jan 21 18:01:33 crc kubenswrapper[4808]: I0121 18:01:33.671707 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tj8vd" event={"ID":"afc0321c-4482-40a4-92e3-67f5f78f5c6e","Type":"ContainerStarted","Data":"f0ea2fe95e73bfa4c9e3423e7f55e70ac2220bd9458d8fc65fb38105aa2d7059"} Jan 21 18:01:33 crc kubenswrapper[4808]: I0121 18:01:33.676853 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59dvj" event={"ID":"bc01e22f-4eb3-46d9-8bba-faffce8f294e","Type":"ContainerStarted","Data":"b7baafb6c0f0a362c33b6c106da6dfcfa2cb8804aa7d26cca1a5488208810b81"} Jan 21 18:01:33 crc kubenswrapper[4808]: I0121 18:01:33.736567 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-59dvj" podStartSLOduration=2.107859858 podStartE2EDuration="4.736552535s" podCreationTimestamp="2026-01-21 18:01:29 +0000 UTC" firstStartedPulling="2026-01-21 18:01:30.634626261 +0000 UTC m=+325.365137146" lastFinishedPulling="2026-01-21 18:01:33.263318938 +0000 UTC m=+327.993829823" observedRunningTime="2026-01-21 18:01:33.736442632 +0000 UTC m=+328.466953527" watchObservedRunningTime="2026-01-21 18:01:33.736552535 +0000 UTC m=+328.467063420" Jan 21 18:01:33 crc kubenswrapper[4808]: I0121 18:01:33.962277 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4pfvs"] Jan 21 18:01:33 crc kubenswrapper[4808]: I0121 18:01:33.963933 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4pfvs" Jan 21 18:01:33 crc kubenswrapper[4808]: I0121 18:01:33.975620 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsx8z\" (UniqueName: \"kubernetes.io/projected/9d1c03b0-e10f-47ae-841c-153e4ae0f9cd-kube-api-access-gsx8z\") pod \"certified-operators-4pfvs\" (UID: \"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd\") " pod="openshift-marketplace/certified-operators-4pfvs" Jan 21 18:01:33 crc kubenswrapper[4808]: I0121 18:01:33.975687 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d1c03b0-e10f-47ae-841c-153e4ae0f9cd-utilities\") pod \"certified-operators-4pfvs\" (UID: \"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd\") " pod="openshift-marketplace/certified-operators-4pfvs" Jan 21 18:01:33 crc kubenswrapper[4808]: I0121 18:01:33.975785 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d1c03b0-e10f-47ae-841c-153e4ae0f9cd-catalog-content\") pod \"certified-operators-4pfvs\" (UID: \"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd\") " pod="openshift-marketplace/certified-operators-4pfvs" Jan 21 18:01:33 crc kubenswrapper[4808]: I0121 18:01:33.979566 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4pfvs"] Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.077130 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d1c03b0-e10f-47ae-841c-153e4ae0f9cd-catalog-content\") pod \"certified-operators-4pfvs\" (UID: \"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd\") " pod="openshift-marketplace/certified-operators-4pfvs" Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.077560 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsx8z\" (UniqueName: \"kubernetes.io/projected/9d1c03b0-e10f-47ae-841c-153e4ae0f9cd-kube-api-access-gsx8z\") pod \"certified-operators-4pfvs\" (UID: \"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd\") " pod="openshift-marketplace/certified-operators-4pfvs" Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.077691 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d1c03b0-e10f-47ae-841c-153e4ae0f9cd-utilities\") pod \"certified-operators-4pfvs\" (UID: \"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd\") " pod="openshift-marketplace/certified-operators-4pfvs" Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.077870 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d1c03b0-e10f-47ae-841c-153e4ae0f9cd-catalog-content\") pod \"certified-operators-4pfvs\" (UID: \"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd\") " pod="openshift-marketplace/certified-operators-4pfvs" Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.078141 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d1c03b0-e10f-47ae-841c-153e4ae0f9cd-utilities\") pod \"certified-operators-4pfvs\" (UID: \"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd\") " pod="openshift-marketplace/certified-operators-4pfvs" Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.103308 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsx8z\" (UniqueName: \"kubernetes.io/projected/9d1c03b0-e10f-47ae-841c-153e4ae0f9cd-kube-api-access-gsx8z\") pod \"certified-operators-4pfvs\" (UID: \"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd\") " pod="openshift-marketplace/certified-operators-4pfvs" Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.303370 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4pfvs" Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.359228 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-74996fc445-x29kx"] Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.359790 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" podUID="961a0fb2-cce4-441d-ab82-37f78562173e" containerName="controller-manager" containerID="cri-o://80c2a38df2887b274a574f57018ed4593139f0eecb763fd1222a3aeb0efa7e79" gracePeriod=30 Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.372953 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6"] Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.373494 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" podUID="8ab26c4b-4728-4cd2-b215-e12f00f396bc" containerName="route-controller-manager" containerID="cri-o://d20814d937d700e2dba9a2b559a2a4cc4fad036d4c288d6bff4d89d17a7f1df0" gracePeriod=30 Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.571878 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hvw6j"] Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.573535 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hvw6j" Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.608363 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hvw6j"] Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.683308 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzw75\" (UniqueName: \"kubernetes.io/projected/e892666d-d0e9-4169-98d7-70687ac29c1b-kube-api-access-qzw75\") pod \"redhat-operators-hvw6j\" (UID: \"e892666d-d0e9-4169-98d7-70687ac29c1b\") " pod="openshift-marketplace/redhat-operators-hvw6j" Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.683369 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e892666d-d0e9-4169-98d7-70687ac29c1b-catalog-content\") pod \"redhat-operators-hvw6j\" (UID: \"e892666d-d0e9-4169-98d7-70687ac29c1b\") " pod="openshift-marketplace/redhat-operators-hvw6j" Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.683393 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e892666d-d0e9-4169-98d7-70687ac29c1b-utilities\") pod \"redhat-operators-hvw6j\" (UID: \"e892666d-d0e9-4169-98d7-70687ac29c1b\") " pod="openshift-marketplace/redhat-operators-hvw6j" Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.683817 4808 generic.go:334] "Generic (PLEG): container finished" podID="961a0fb2-cce4-441d-ab82-37f78562173e" containerID="80c2a38df2887b274a574f57018ed4593139f0eecb763fd1222a3aeb0efa7e79" exitCode=0 Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.683874 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" event={"ID":"961a0fb2-cce4-441d-ab82-37f78562173e","Type":"ContainerDied","Data":"80c2a38df2887b274a574f57018ed4593139f0eecb763fd1222a3aeb0efa7e79"} Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.686920 4808 generic.go:334] "Generic (PLEG): container finished" podID="8ab26c4b-4728-4cd2-b215-e12f00f396bc" containerID="d20814d937d700e2dba9a2b559a2a4cc4fad036d4c288d6bff4d89d17a7f1df0" exitCode=0 Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.686985 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" event={"ID":"8ab26c4b-4728-4cd2-b215-e12f00f396bc","Type":"ContainerDied","Data":"d20814d937d700e2dba9a2b559a2a4cc4fad036d4c288d6bff4d89d17a7f1df0"} Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.693667 4808 generic.go:334] "Generic (PLEG): container finished" podID="173b6b54-1024-4806-8056-985010126525" containerID="5202c43b50f1265bbd8a3ed0aa836614a8f0ddc4a82081f921e026f26dd76b02" exitCode=0 Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.694421 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xstv7" event={"ID":"173b6b54-1024-4806-8056-985010126525","Type":"ContainerDied","Data":"5202c43b50f1265bbd8a3ed0aa836614a8f0ddc4a82081f921e026f26dd76b02"} Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.786941 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e892666d-d0e9-4169-98d7-70687ac29c1b-utilities\") pod \"redhat-operators-hvw6j\" (UID: \"e892666d-d0e9-4169-98d7-70687ac29c1b\") " pod="openshift-marketplace/redhat-operators-hvw6j" Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.787347 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzw75\" (UniqueName: \"kubernetes.io/projected/e892666d-d0e9-4169-98d7-70687ac29c1b-kube-api-access-qzw75\") pod \"redhat-operators-hvw6j\" (UID: \"e892666d-d0e9-4169-98d7-70687ac29c1b\") " pod="openshift-marketplace/redhat-operators-hvw6j" Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.787386 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e892666d-d0e9-4169-98d7-70687ac29c1b-catalog-content\") pod \"redhat-operators-hvw6j\" (UID: \"e892666d-d0e9-4169-98d7-70687ac29c1b\") " pod="openshift-marketplace/redhat-operators-hvw6j" Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.788319 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e892666d-d0e9-4169-98d7-70687ac29c1b-catalog-content\") pod \"redhat-operators-hvw6j\" (UID: \"e892666d-d0e9-4169-98d7-70687ac29c1b\") " pod="openshift-marketplace/redhat-operators-hvw6j" Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.788315 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e892666d-d0e9-4169-98d7-70687ac29c1b-utilities\") pod \"redhat-operators-hvw6j\" (UID: \"e892666d-d0e9-4169-98d7-70687ac29c1b\") " pod="openshift-marketplace/redhat-operators-hvw6j" Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.807870 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzw75\" (UniqueName: \"kubernetes.io/projected/e892666d-d0e9-4169-98d7-70687ac29c1b-kube-api-access-qzw75\") pod \"redhat-operators-hvw6j\" (UID: \"e892666d-d0e9-4169-98d7-70687ac29c1b\") " pod="openshift-marketplace/redhat-operators-hvw6j" Jan 21 18:01:34 crc kubenswrapper[4808]: I0121 18:01:34.893925 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hvw6j" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.099361 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6m94d" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.100816 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6m94d" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.160995 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4pfvs"] Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.195849 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6m94d" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.244600 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hvw6j"] Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.292306 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.297579 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fp7h8\" (UniqueName: \"kubernetes.io/projected/8ab26c4b-4728-4cd2-b215-e12f00f396bc-kube-api-access-fp7h8\") pod \"8ab26c4b-4728-4cd2-b215-e12f00f396bc\" (UID: \"8ab26c4b-4728-4cd2-b215-e12f00f396bc\") " Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.297648 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ab26c4b-4728-4cd2-b215-e12f00f396bc-config\") pod \"8ab26c4b-4728-4cd2-b215-e12f00f396bc\" (UID: \"8ab26c4b-4728-4cd2-b215-e12f00f396bc\") " Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.297676 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ab26c4b-4728-4cd2-b215-e12f00f396bc-serving-cert\") pod \"8ab26c4b-4728-4cd2-b215-e12f00f396bc\" (UID: \"8ab26c4b-4728-4cd2-b215-e12f00f396bc\") " Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.297741 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ab26c4b-4728-4cd2-b215-e12f00f396bc-client-ca\") pod \"8ab26c4b-4728-4cd2-b215-e12f00f396bc\" (UID: \"8ab26c4b-4728-4cd2-b215-e12f00f396bc\") " Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.304230 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ab26c4b-4728-4cd2-b215-e12f00f396bc-client-ca" (OuterVolumeSpecName: "client-ca") pod "8ab26c4b-4728-4cd2-b215-e12f00f396bc" (UID: "8ab26c4b-4728-4cd2-b215-e12f00f396bc"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.304955 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ab26c4b-4728-4cd2-b215-e12f00f396bc-config" (OuterVolumeSpecName: "config") pod "8ab26c4b-4728-4cd2-b215-e12f00f396bc" (UID: "8ab26c4b-4728-4cd2-b215-e12f00f396bc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.317501 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ab26c4b-4728-4cd2-b215-e12f00f396bc-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8ab26c4b-4728-4cd2-b215-e12f00f396bc" (UID: "8ab26c4b-4728-4cd2-b215-e12f00f396bc"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.317495 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ab26c4b-4728-4cd2-b215-e12f00f396bc-kube-api-access-fp7h8" (OuterVolumeSpecName: "kube-api-access-fp7h8") pod "8ab26c4b-4728-4cd2-b215-e12f00f396bc" (UID: "8ab26c4b-4728-4cd2-b215-e12f00f396bc"). InnerVolumeSpecName "kube-api-access-fp7h8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.386434 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.398879 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ab26c4b-4728-4cd2-b215-e12f00f396bc-config\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.398926 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ab26c4b-4728-4cd2-b215-e12f00f396bc-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.398941 4808 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ab26c4b-4728-4cd2-b215-e12f00f396bc-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.398954 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fp7h8\" (UniqueName: \"kubernetes.io/projected/8ab26c4b-4728-4cd2-b215-e12f00f396bc-kube-api-access-fp7h8\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.499625 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/961a0fb2-cce4-441d-ab82-37f78562173e-config\") pod \"961a0fb2-cce4-441d-ab82-37f78562173e\" (UID: \"961a0fb2-cce4-441d-ab82-37f78562173e\") " Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.500028 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phvss\" (UniqueName: \"kubernetes.io/projected/961a0fb2-cce4-441d-ab82-37f78562173e-kube-api-access-phvss\") pod \"961a0fb2-cce4-441d-ab82-37f78562173e\" (UID: \"961a0fb2-cce4-441d-ab82-37f78562173e\") " Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.500057 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/961a0fb2-cce4-441d-ab82-37f78562173e-client-ca\") pod \"961a0fb2-cce4-441d-ab82-37f78562173e\" (UID: \"961a0fb2-cce4-441d-ab82-37f78562173e\") " Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.500138 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/961a0fb2-cce4-441d-ab82-37f78562173e-serving-cert\") pod \"961a0fb2-cce4-441d-ab82-37f78562173e\" (UID: \"961a0fb2-cce4-441d-ab82-37f78562173e\") " Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.500192 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/961a0fb2-cce4-441d-ab82-37f78562173e-proxy-ca-bundles\") pod \"961a0fb2-cce4-441d-ab82-37f78562173e\" (UID: \"961a0fb2-cce4-441d-ab82-37f78562173e\") " Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.501304 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/961a0fb2-cce4-441d-ab82-37f78562173e-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "961a0fb2-cce4-441d-ab82-37f78562173e" (UID: "961a0fb2-cce4-441d-ab82-37f78562173e"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.501975 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/961a0fb2-cce4-441d-ab82-37f78562173e-client-ca" (OuterVolumeSpecName: "client-ca") pod "961a0fb2-cce4-441d-ab82-37f78562173e" (UID: "961a0fb2-cce4-441d-ab82-37f78562173e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.502420 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/961a0fb2-cce4-441d-ab82-37f78562173e-config" (OuterVolumeSpecName: "config") pod "961a0fb2-cce4-441d-ab82-37f78562173e" (UID: "961a0fb2-cce4-441d-ab82-37f78562173e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.505983 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/961a0fb2-cce4-441d-ab82-37f78562173e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "961a0fb2-cce4-441d-ab82-37f78562173e" (UID: "961a0fb2-cce4-441d-ab82-37f78562173e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.511416 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/961a0fb2-cce4-441d-ab82-37f78562173e-kube-api-access-phvss" (OuterVolumeSpecName: "kube-api-access-phvss") pod "961a0fb2-cce4-441d-ab82-37f78562173e" (UID: "961a0fb2-cce4-441d-ab82-37f78562173e"). InnerVolumeSpecName "kube-api-access-phvss". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.601890 4808 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/961a0fb2-cce4-441d-ab82-37f78562173e-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.601922 4808 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/961a0fb2-cce4-441d-ab82-37f78562173e-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.601933 4808 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/961a0fb2-cce4-441d-ab82-37f78562173e-config\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.601942 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phvss\" (UniqueName: \"kubernetes.io/projected/961a0fb2-cce4-441d-ab82-37f78562173e-kube-api-access-phvss\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.601951 4808 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/961a0fb2-cce4-441d-ab82-37f78562173e-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.699708 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pfvs" event={"ID":"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd","Type":"ContainerStarted","Data":"099e0d2b82bdcb9221be0dd433b35617d145bbc57a883d09664d085ded5f25a5"} Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.699760 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pfvs" event={"ID":"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd","Type":"ContainerStarted","Data":"a1bc51b30d252ba93ed5610738e6486733cf5b2409100eed98a5ad1a067f6f29"} Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.701006 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.701005 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74996fc445-x29kx" event={"ID":"961a0fb2-cce4-441d-ab82-37f78562173e","Type":"ContainerDied","Data":"0bc465d23218a98c9503a35dca58a781d1278b559569230e625106764d4cbbaf"} Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.701140 4808 scope.go:117] "RemoveContainer" containerID="80c2a38df2887b274a574f57018ed4593139f0eecb763fd1222a3aeb0efa7e79" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.702701 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" event={"ID":"8ab26c4b-4728-4cd2-b215-e12f00f396bc","Type":"ContainerDied","Data":"0e651e60388eb4c23b5fe19f446f9a8bebdd9668c7840a7a4aabb1276f1779c5"} Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.703030 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.705335 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xstv7" event={"ID":"173b6b54-1024-4806-8056-985010126525","Type":"ContainerStarted","Data":"af4daed6d8653d6e8524448329a6f2108446a3f501ec5904015487ff2138977f"} Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.709755 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tj8vd" event={"ID":"afc0321c-4482-40a4-92e3-67f5f78f5c6e","Type":"ContainerStarted","Data":"dd0575dcbea2818c8a92d5c07ae200912c83915e94e89e2022df47d38d5b724d"} Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.715558 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvw6j" event={"ID":"e892666d-d0e9-4169-98d7-70687ac29c1b","Type":"ContainerStarted","Data":"d4a093edd02c8e68c2aa58fe0aa83e7787da67e14f33ffa792dd72dc02c62876"} Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.715602 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvw6j" event={"ID":"e892666d-d0e9-4169-98d7-70687ac29c1b","Type":"ContainerStarted","Data":"8659b08db5ae9b553c676ede57d80747a598f593d5979829ecbcddb403d7b63c"} Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.719859 4808 scope.go:117] "RemoveContainer" containerID="d20814d937d700e2dba9a2b559a2a4cc4fad036d4c288d6bff4d89d17a7f1df0" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.761587 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xstv7" podStartSLOduration=2.23480715 podStartE2EDuration="4.761568427s" podCreationTimestamp="2026-01-21 18:01:31 +0000 UTC" firstStartedPulling="2026-01-21 18:01:32.658553908 +0000 UTC m=+327.389064793" lastFinishedPulling="2026-01-21 18:01:35.185315185 +0000 UTC m=+329.915826070" observedRunningTime="2026-01-21 18:01:35.760281746 +0000 UTC m=+330.490792641" watchObservedRunningTime="2026-01-21 18:01:35.761568427 +0000 UTC m=+330.492079302" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.767182 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6m94d" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.778301 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6"] Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.781525 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-648ff87cd5-7ngm6"] Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.809211 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx"] Jan 21 18:01:35 crc kubenswrapper[4808]: E0121 18:01:35.809470 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="961a0fb2-cce4-441d-ab82-37f78562173e" containerName="controller-manager" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.809484 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="961a0fb2-cce4-441d-ab82-37f78562173e" containerName="controller-manager" Jan 21 18:01:35 crc kubenswrapper[4808]: E0121 18:01:35.809494 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ab26c4b-4728-4cd2-b215-e12f00f396bc" containerName="route-controller-manager" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.809500 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ab26c4b-4728-4cd2-b215-e12f00f396bc" containerName="route-controller-manager" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.809610 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="961a0fb2-cce4-441d-ab82-37f78562173e" containerName="controller-manager" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.809624 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ab26c4b-4728-4cd2-b215-e12f00f396bc" containerName="route-controller-manager" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.810013 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.814797 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.814994 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.815105 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.815579 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.815783 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.821272 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.874326 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-b5559d4f4-jzxjc"] Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.875321 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.877337 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-74996fc445-x29kx"] Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.881430 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.881509 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.881675 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.882891 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.882959 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.883053 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.883077 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx"] Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.896993 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.898333 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-74996fc445-x29kx"] Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.902490 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-b5559d4f4-jzxjc"] Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.905703 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bffe325b-1095-41f0-ae24-6bcffa3ebf90-serving-cert\") pod \"route-controller-manager-7b8568c6bf-hvqvx\" (UID: \"bffe325b-1095-41f0-ae24-6bcffa3ebf90\") " pod="openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.905742 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bffe325b-1095-41f0-ae24-6bcffa3ebf90-config\") pod \"route-controller-manager-7b8568c6bf-hvqvx\" (UID: \"bffe325b-1095-41f0-ae24-6bcffa3ebf90\") " pod="openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.905765 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k4vs\" (UniqueName: \"kubernetes.io/projected/bffe325b-1095-41f0-ae24-6bcffa3ebf90-kube-api-access-2k4vs\") pod \"route-controller-manager-7b8568c6bf-hvqvx\" (UID: \"bffe325b-1095-41f0-ae24-6bcffa3ebf90\") " pod="openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx" Jan 21 18:01:35 crc kubenswrapper[4808]: I0121 18:01:35.905843 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bffe325b-1095-41f0-ae24-6bcffa3ebf90-client-ca\") pod \"route-controller-manager-7b8568c6bf-hvqvx\" (UID: \"bffe325b-1095-41f0-ae24-6bcffa3ebf90\") " pod="openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.007760 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ddb5c43-56f5-42a3-b703-003a5e2fa448-config\") pod \"controller-manager-b5559d4f4-jzxjc\" (UID: \"9ddb5c43-56f5-42a3-b703-003a5e2fa448\") " pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.008270 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bffe325b-1095-41f0-ae24-6bcffa3ebf90-serving-cert\") pod \"route-controller-manager-7b8568c6bf-hvqvx\" (UID: \"bffe325b-1095-41f0-ae24-6bcffa3ebf90\") " pod="openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.008307 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ddb5c43-56f5-42a3-b703-003a5e2fa448-client-ca\") pod \"controller-manager-b5559d4f4-jzxjc\" (UID: \"9ddb5c43-56f5-42a3-b703-003a5e2fa448\") " pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.008352 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvppx\" (UniqueName: \"kubernetes.io/projected/9ddb5c43-56f5-42a3-b703-003a5e2fa448-kube-api-access-hvppx\") pod \"controller-manager-b5559d4f4-jzxjc\" (UID: \"9ddb5c43-56f5-42a3-b703-003a5e2fa448\") " pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.008385 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bffe325b-1095-41f0-ae24-6bcffa3ebf90-config\") pod \"route-controller-manager-7b8568c6bf-hvqvx\" (UID: \"bffe325b-1095-41f0-ae24-6bcffa3ebf90\") " pod="openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.008442 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k4vs\" (UniqueName: \"kubernetes.io/projected/bffe325b-1095-41f0-ae24-6bcffa3ebf90-kube-api-access-2k4vs\") pod \"route-controller-manager-7b8568c6bf-hvqvx\" (UID: \"bffe325b-1095-41f0-ae24-6bcffa3ebf90\") " pod="openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.008534 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9ddb5c43-56f5-42a3-b703-003a5e2fa448-proxy-ca-bundles\") pod \"controller-manager-b5559d4f4-jzxjc\" (UID: \"9ddb5c43-56f5-42a3-b703-003a5e2fa448\") " pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.008597 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ddb5c43-56f5-42a3-b703-003a5e2fa448-serving-cert\") pod \"controller-manager-b5559d4f4-jzxjc\" (UID: \"9ddb5c43-56f5-42a3-b703-003a5e2fa448\") " pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.008628 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bffe325b-1095-41f0-ae24-6bcffa3ebf90-client-ca\") pod \"route-controller-manager-7b8568c6bf-hvqvx\" (UID: \"bffe325b-1095-41f0-ae24-6bcffa3ebf90\") " pod="openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.010517 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bffe325b-1095-41f0-ae24-6bcffa3ebf90-config\") pod \"route-controller-manager-7b8568c6bf-hvqvx\" (UID: \"bffe325b-1095-41f0-ae24-6bcffa3ebf90\") " pod="openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.010581 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bffe325b-1095-41f0-ae24-6bcffa3ebf90-client-ca\") pod \"route-controller-manager-7b8568c6bf-hvqvx\" (UID: \"bffe325b-1095-41f0-ae24-6bcffa3ebf90\") " pod="openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.016000 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bffe325b-1095-41f0-ae24-6bcffa3ebf90-serving-cert\") pod \"route-controller-manager-7b8568c6bf-hvqvx\" (UID: \"bffe325b-1095-41f0-ae24-6bcffa3ebf90\") " pod="openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.032785 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k4vs\" (UniqueName: \"kubernetes.io/projected/bffe325b-1095-41f0-ae24-6bcffa3ebf90-kube-api-access-2k4vs\") pod \"route-controller-manager-7b8568c6bf-hvqvx\" (UID: \"bffe325b-1095-41f0-ae24-6bcffa3ebf90\") " pod="openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.110610 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9ddb5c43-56f5-42a3-b703-003a5e2fa448-proxy-ca-bundles\") pod \"controller-manager-b5559d4f4-jzxjc\" (UID: \"9ddb5c43-56f5-42a3-b703-003a5e2fa448\") " pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.110716 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ddb5c43-56f5-42a3-b703-003a5e2fa448-serving-cert\") pod \"controller-manager-b5559d4f4-jzxjc\" (UID: \"9ddb5c43-56f5-42a3-b703-003a5e2fa448\") " pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.110761 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ddb5c43-56f5-42a3-b703-003a5e2fa448-config\") pod \"controller-manager-b5559d4f4-jzxjc\" (UID: \"9ddb5c43-56f5-42a3-b703-003a5e2fa448\") " pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.110788 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ddb5c43-56f5-42a3-b703-003a5e2fa448-client-ca\") pod \"controller-manager-b5559d4f4-jzxjc\" (UID: \"9ddb5c43-56f5-42a3-b703-003a5e2fa448\") " pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.110806 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvppx\" (UniqueName: \"kubernetes.io/projected/9ddb5c43-56f5-42a3-b703-003a5e2fa448-kube-api-access-hvppx\") pod \"controller-manager-b5559d4f4-jzxjc\" (UID: \"9ddb5c43-56f5-42a3-b703-003a5e2fa448\") " pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.112183 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ddb5c43-56f5-42a3-b703-003a5e2fa448-client-ca\") pod \"controller-manager-b5559d4f4-jzxjc\" (UID: \"9ddb5c43-56f5-42a3-b703-003a5e2fa448\") " pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.113149 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9ddb5c43-56f5-42a3-b703-003a5e2fa448-proxy-ca-bundles\") pod \"controller-manager-b5559d4f4-jzxjc\" (UID: \"9ddb5c43-56f5-42a3-b703-003a5e2fa448\") " pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.113495 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ddb5c43-56f5-42a3-b703-003a5e2fa448-config\") pod \"controller-manager-b5559d4f4-jzxjc\" (UID: \"9ddb5c43-56f5-42a3-b703-003a5e2fa448\") " pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.118556 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ddb5c43-56f5-42a3-b703-003a5e2fa448-serving-cert\") pod \"controller-manager-b5559d4f4-jzxjc\" (UID: \"9ddb5c43-56f5-42a3-b703-003a5e2fa448\") " pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.130088 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvppx\" (UniqueName: \"kubernetes.io/projected/9ddb5c43-56f5-42a3-b703-003a5e2fa448-kube-api-access-hvppx\") pod \"controller-manager-b5559d4f4-jzxjc\" (UID: \"9ddb5c43-56f5-42a3-b703-003a5e2fa448\") " pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.206376 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.250133 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.369687 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cbwpb"] Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.371187 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cbwpb" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.404802 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cbwpb"] Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.504799 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx"] Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.515918 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mv79m\" (UniqueName: \"kubernetes.io/projected/1df03daa-d900-4a6d-9776-26f3b05843ce-kube-api-access-mv79m\") pod \"certified-operators-cbwpb\" (UID: \"1df03daa-d900-4a6d-9776-26f3b05843ce\") " pod="openshift-marketplace/certified-operators-cbwpb" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.515987 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df03daa-d900-4a6d-9776-26f3b05843ce-utilities\") pod \"certified-operators-cbwpb\" (UID: \"1df03daa-d900-4a6d-9776-26f3b05843ce\") " pod="openshift-marketplace/certified-operators-cbwpb" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.516025 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df03daa-d900-4a6d-9776-26f3b05843ce-catalog-content\") pod \"certified-operators-cbwpb\" (UID: \"1df03daa-d900-4a6d-9776-26f3b05843ce\") " pod="openshift-marketplace/certified-operators-cbwpb" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.594849 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-b5559d4f4-jzxjc"] Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.617088 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df03daa-d900-4a6d-9776-26f3b05843ce-catalog-content\") pod \"certified-operators-cbwpb\" (UID: \"1df03daa-d900-4a6d-9776-26f3b05843ce\") " pod="openshift-marketplace/certified-operators-cbwpb" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.617181 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mv79m\" (UniqueName: \"kubernetes.io/projected/1df03daa-d900-4a6d-9776-26f3b05843ce-kube-api-access-mv79m\") pod \"certified-operators-cbwpb\" (UID: \"1df03daa-d900-4a6d-9776-26f3b05843ce\") " pod="openshift-marketplace/certified-operators-cbwpb" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.617229 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df03daa-d900-4a6d-9776-26f3b05843ce-utilities\") pod \"certified-operators-cbwpb\" (UID: \"1df03daa-d900-4a6d-9776-26f3b05843ce\") " pod="openshift-marketplace/certified-operators-cbwpb" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.617731 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df03daa-d900-4a6d-9776-26f3b05843ce-catalog-content\") pod \"certified-operators-cbwpb\" (UID: \"1df03daa-d900-4a6d-9776-26f3b05843ce\") " pod="openshift-marketplace/certified-operators-cbwpb" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.619624 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df03daa-d900-4a6d-9776-26f3b05843ce-utilities\") pod \"certified-operators-cbwpb\" (UID: \"1df03daa-d900-4a6d-9776-26f3b05843ce\") " pod="openshift-marketplace/certified-operators-cbwpb" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.639944 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mv79m\" (UniqueName: \"kubernetes.io/projected/1df03daa-d900-4a6d-9776-26f3b05843ce-kube-api-access-mv79m\") pod \"certified-operators-cbwpb\" (UID: \"1df03daa-d900-4a6d-9776-26f3b05843ce\") " pod="openshift-marketplace/certified-operators-cbwpb" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.697631 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cbwpb" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.720989 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx" event={"ID":"bffe325b-1095-41f0-ae24-6bcffa3ebf90","Type":"ContainerStarted","Data":"c67f5ac9c96dac1e76d13a4ef15b4d902d19fe6874f28cbdcafe91951b4d007d"} Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.969216 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wbqf6"] Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.970404 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wbqf6" Jan 21 18:01:36 crc kubenswrapper[4808]: I0121 18:01:36.991559 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wbqf6"] Jan 21 18:01:37 crc kubenswrapper[4808]: I0121 18:01:37.126146 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56xw4\" (UniqueName: \"kubernetes.io/projected/37705e9c-abf0-42ac-8fc3-03e0b2fa732a-kube-api-access-56xw4\") pod \"redhat-operators-wbqf6\" (UID: \"37705e9c-abf0-42ac-8fc3-03e0b2fa732a\") " pod="openshift-marketplace/redhat-operators-wbqf6" Jan 21 18:01:37 crc kubenswrapper[4808]: I0121 18:01:37.126620 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37705e9c-abf0-42ac-8fc3-03e0b2fa732a-catalog-content\") pod \"redhat-operators-wbqf6\" (UID: \"37705e9c-abf0-42ac-8fc3-03e0b2fa732a\") " pod="openshift-marketplace/redhat-operators-wbqf6" Jan 21 18:01:37 crc kubenswrapper[4808]: I0121 18:01:37.126659 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37705e9c-abf0-42ac-8fc3-03e0b2fa732a-utilities\") pod \"redhat-operators-wbqf6\" (UID: \"37705e9c-abf0-42ac-8fc3-03e0b2fa732a\") " pod="openshift-marketplace/redhat-operators-wbqf6" Jan 21 18:01:37 crc kubenswrapper[4808]: W0121 18:01:37.145769 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ddb5c43_56f5_42a3_b703_003a5e2fa448.slice/crio-835d19cb8bbb46bd5e745e76e95330f305115dc540a4b3dfc54564dccdfcad8a WatchSource:0}: Error finding container 835d19cb8bbb46bd5e745e76e95330f305115dc540a4b3dfc54564dccdfcad8a: Status 404 returned error can't find the container with id 835d19cb8bbb46bd5e745e76e95330f305115dc540a4b3dfc54564dccdfcad8a Jan 21 18:01:37 crc kubenswrapper[4808]: I0121 18:01:37.228418 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56xw4\" (UniqueName: \"kubernetes.io/projected/37705e9c-abf0-42ac-8fc3-03e0b2fa732a-kube-api-access-56xw4\") pod \"redhat-operators-wbqf6\" (UID: \"37705e9c-abf0-42ac-8fc3-03e0b2fa732a\") " pod="openshift-marketplace/redhat-operators-wbqf6" Jan 21 18:01:37 crc kubenswrapper[4808]: I0121 18:01:37.228521 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37705e9c-abf0-42ac-8fc3-03e0b2fa732a-catalog-content\") pod \"redhat-operators-wbqf6\" (UID: \"37705e9c-abf0-42ac-8fc3-03e0b2fa732a\") " pod="openshift-marketplace/redhat-operators-wbqf6" Jan 21 18:01:37 crc kubenswrapper[4808]: I0121 18:01:37.228554 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37705e9c-abf0-42ac-8fc3-03e0b2fa732a-utilities\") pod \"redhat-operators-wbqf6\" (UID: \"37705e9c-abf0-42ac-8fc3-03e0b2fa732a\") " pod="openshift-marketplace/redhat-operators-wbqf6" Jan 21 18:01:37 crc kubenswrapper[4808]: I0121 18:01:37.229029 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37705e9c-abf0-42ac-8fc3-03e0b2fa732a-catalog-content\") pod \"redhat-operators-wbqf6\" (UID: \"37705e9c-abf0-42ac-8fc3-03e0b2fa732a\") " pod="openshift-marketplace/redhat-operators-wbqf6" Jan 21 18:01:37 crc kubenswrapper[4808]: I0121 18:01:37.229083 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37705e9c-abf0-42ac-8fc3-03e0b2fa732a-utilities\") pod \"redhat-operators-wbqf6\" (UID: \"37705e9c-abf0-42ac-8fc3-03e0b2fa732a\") " pod="openshift-marketplace/redhat-operators-wbqf6" Jan 21 18:01:37 crc kubenswrapper[4808]: I0121 18:01:37.257361 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56xw4\" (UniqueName: \"kubernetes.io/projected/37705e9c-abf0-42ac-8fc3-03e0b2fa732a-kube-api-access-56xw4\") pod \"redhat-operators-wbqf6\" (UID: \"37705e9c-abf0-42ac-8fc3-03e0b2fa732a\") " pod="openshift-marketplace/redhat-operators-wbqf6" Jan 21 18:01:37 crc kubenswrapper[4808]: I0121 18:01:37.288297 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wbqf6" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:37.434581 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cbwpb"] Jan 21 18:01:39 crc kubenswrapper[4808]: W0121 18:01:37.465618 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1df03daa_d900_4a6d_9776_26f3b05843ce.slice/crio-341be48dd7e1672a511bd2cbc7dc1fc96983b7b7430fd43553f983fb8b68d9fd WatchSource:0}: Error finding container 341be48dd7e1672a511bd2cbc7dc1fc96983b7b7430fd43553f983fb8b68d9fd: Status 404 returned error can't find the container with id 341be48dd7e1672a511bd2cbc7dc1fc96983b7b7430fd43553f983fb8b68d9fd Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:37.481191 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-54shl" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:37.481235 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-54shl" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:37.528704 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ab26c4b-4728-4cd2-b215-e12f00f396bc" path="/var/lib/kubelet/pods/8ab26c4b-4728-4cd2-b215-e12f00f396bc/volumes" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:37.529526 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="961a0fb2-cce4-441d-ab82-37f78562173e" path="/var/lib/kubelet/pods/961a0fb2-cce4-441d-ab82-37f78562173e/volumes" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:37.728858 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" event={"ID":"9ddb5c43-56f5-42a3-b703-003a5e2fa448","Type":"ContainerStarted","Data":"f232b050952fe3d3e0cf1a92d68e04fc2d9be4cb0d612cae6df58e152156c5c6"} Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:37.728905 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" event={"ID":"9ddb5c43-56f5-42a3-b703-003a5e2fa448","Type":"ContainerStarted","Data":"835d19cb8bbb46bd5e745e76e95330f305115dc540a4b3dfc54564dccdfcad8a"} Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:37.730609 4808 generic.go:334] "Generic (PLEG): container finished" podID="9d1c03b0-e10f-47ae-841c-153e4ae0f9cd" containerID="099e0d2b82bdcb9221be0dd433b35617d145bbc57a883d09664d085ded5f25a5" exitCode=0 Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:37.730659 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pfvs" event={"ID":"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd","Type":"ContainerDied","Data":"099e0d2b82bdcb9221be0dd433b35617d145bbc57a883d09664d085ded5f25a5"} Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:37.733526 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx" event={"ID":"bffe325b-1095-41f0-ae24-6bcffa3ebf90","Type":"ContainerStarted","Data":"ab8ea14b3b4692d3ca17c5d77501fef553a36e040479f4ae56166820f29484f8"} Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:37.733945 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:37.737076 4808 generic.go:334] "Generic (PLEG): container finished" podID="e892666d-d0e9-4169-98d7-70687ac29c1b" containerID="d4a093edd02c8e68c2aa58fe0aa83e7787da67e14f33ffa792dd72dc02c62876" exitCode=0 Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:37.737152 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvw6j" event={"ID":"e892666d-d0e9-4169-98d7-70687ac29c1b","Type":"ContainerDied","Data":"d4a093edd02c8e68c2aa58fe0aa83e7787da67e14f33ffa792dd72dc02c62876"} Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:37.742096 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cbwpb" event={"ID":"1df03daa-d900-4a6d-9776-26f3b05843ce","Type":"ContainerStarted","Data":"341be48dd7e1672a511bd2cbc7dc1fc96983b7b7430fd43553f983fb8b68d9fd"} Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:37.793460 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx" podStartSLOduration=3.793427108 podStartE2EDuration="3.793427108s" podCreationTimestamp="2026-01-21 18:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 18:01:37.791257785 +0000 UTC m=+332.521768720" watchObservedRunningTime="2026-01-21 18:01:37.793427108 +0000 UTC m=+332.523937993" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:38.273717 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7b8568c6bf-hvqvx" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:38.549905 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-54shl" podUID="9c4c6107-02cd-47e8-b29a-88190d15f030" containerName="registry-server" probeResult="failure" output=< Jan 21 18:01:39 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Jan 21 18:01:39 crc kubenswrapper[4808]: > Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:38.749807 4808 generic.go:334] "Generic (PLEG): container finished" podID="afc0321c-4482-40a4-92e3-67f5f78f5c6e" containerID="dd0575dcbea2818c8a92d5c07ae200912c83915e94e89e2022df47d38d5b724d" exitCode=0 Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:38.749875 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tj8vd" event={"ID":"afc0321c-4482-40a4-92e3-67f5f78f5c6e","Type":"ContainerDied","Data":"dd0575dcbea2818c8a92d5c07ae200912c83915e94e89e2022df47d38d5b724d"} Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:38.757622 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cbwpb" event={"ID":"1df03daa-d900-4a6d-9776-26f3b05843ce","Type":"ContainerStarted","Data":"4d2be47b39e77a53c6b82949dfa1368760aa914be579e7af148f4593039c2521"} Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:38.758268 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:38.767924 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:38.785403 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5c9xs"] Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:38.786756 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5c9xs" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:38.794940 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5c9xs"] Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:38.828050 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-b5559d4f4-jzxjc" podStartSLOduration=4.828031544 podStartE2EDuration="4.828031544s" podCreationTimestamp="2026-01-21 18:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 18:01:38.827513861 +0000 UTC m=+333.558024756" watchObservedRunningTime="2026-01-21 18:01:38.828031544 +0000 UTC m=+333.558542419" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:38.953320 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p69s2\" (UniqueName: \"kubernetes.io/projected/21ebff69-0bfe-427d-8330-4e454159dd01-kube-api-access-p69s2\") pod \"certified-operators-5c9xs\" (UID: \"21ebff69-0bfe-427d-8330-4e454159dd01\") " pod="openshift-marketplace/certified-operators-5c9xs" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:38.953392 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21ebff69-0bfe-427d-8330-4e454159dd01-utilities\") pod \"certified-operators-5c9xs\" (UID: \"21ebff69-0bfe-427d-8330-4e454159dd01\") " pod="openshift-marketplace/certified-operators-5c9xs" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:38.953440 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21ebff69-0bfe-427d-8330-4e454159dd01-catalog-content\") pod \"certified-operators-5c9xs\" (UID: \"21ebff69-0bfe-427d-8330-4e454159dd01\") " pod="openshift-marketplace/certified-operators-5c9xs" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.055478 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21ebff69-0bfe-427d-8330-4e454159dd01-utilities\") pod \"certified-operators-5c9xs\" (UID: \"21ebff69-0bfe-427d-8330-4e454159dd01\") " pod="openshift-marketplace/certified-operators-5c9xs" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.055559 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21ebff69-0bfe-427d-8330-4e454159dd01-catalog-content\") pod \"certified-operators-5c9xs\" (UID: \"21ebff69-0bfe-427d-8330-4e454159dd01\") " pod="openshift-marketplace/certified-operators-5c9xs" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.055617 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p69s2\" (UniqueName: \"kubernetes.io/projected/21ebff69-0bfe-427d-8330-4e454159dd01-kube-api-access-p69s2\") pod \"certified-operators-5c9xs\" (UID: \"21ebff69-0bfe-427d-8330-4e454159dd01\") " pod="openshift-marketplace/certified-operators-5c9xs" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.056364 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21ebff69-0bfe-427d-8330-4e454159dd01-utilities\") pod \"certified-operators-5c9xs\" (UID: \"21ebff69-0bfe-427d-8330-4e454159dd01\") " pod="openshift-marketplace/certified-operators-5c9xs" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.056624 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21ebff69-0bfe-427d-8330-4e454159dd01-catalog-content\") pod \"certified-operators-5c9xs\" (UID: \"21ebff69-0bfe-427d-8330-4e454159dd01\") " pod="openshift-marketplace/certified-operators-5c9xs" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.083406 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p69s2\" (UniqueName: \"kubernetes.io/projected/21ebff69-0bfe-427d-8330-4e454159dd01-kube-api-access-p69s2\") pod \"certified-operators-5c9xs\" (UID: \"21ebff69-0bfe-427d-8330-4e454159dd01\") " pod="openshift-marketplace/certified-operators-5c9xs" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.274690 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5c9xs" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.334718 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hzsq2" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.334763 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hzsq2" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.371376 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jl45c"] Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.372677 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jl45c" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.384734 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hzsq2" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.393230 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jl45c"] Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.460101 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbcgm\" (UniqueName: \"kubernetes.io/projected/ff29d64c-c8c9-4cb7-9a58-4f04f78034e2-kube-api-access-qbcgm\") pod \"redhat-operators-jl45c\" (UID: \"ff29d64c-c8c9-4cb7-9a58-4f04f78034e2\") " pod="openshift-marketplace/redhat-operators-jl45c" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.460149 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff29d64c-c8c9-4cb7-9a58-4f04f78034e2-catalog-content\") pod \"redhat-operators-jl45c\" (UID: \"ff29d64c-c8c9-4cb7-9a58-4f04f78034e2\") " pod="openshift-marketplace/redhat-operators-jl45c" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.460174 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff29d64c-c8c9-4cb7-9a58-4f04f78034e2-utilities\") pod \"redhat-operators-jl45c\" (UID: \"ff29d64c-c8c9-4cb7-9a58-4f04f78034e2\") " pod="openshift-marketplace/redhat-operators-jl45c" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.561215 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbcgm\" (UniqueName: \"kubernetes.io/projected/ff29d64c-c8c9-4cb7-9a58-4f04f78034e2-kube-api-access-qbcgm\") pod \"redhat-operators-jl45c\" (UID: \"ff29d64c-c8c9-4cb7-9a58-4f04f78034e2\") " pod="openshift-marketplace/redhat-operators-jl45c" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.561286 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff29d64c-c8c9-4cb7-9a58-4f04f78034e2-catalog-content\") pod \"redhat-operators-jl45c\" (UID: \"ff29d64c-c8c9-4cb7-9a58-4f04f78034e2\") " pod="openshift-marketplace/redhat-operators-jl45c" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.561311 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff29d64c-c8c9-4cb7-9a58-4f04f78034e2-utilities\") pod \"redhat-operators-jl45c\" (UID: \"ff29d64c-c8c9-4cb7-9a58-4f04f78034e2\") " pod="openshift-marketplace/redhat-operators-jl45c" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.561791 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff29d64c-c8c9-4cb7-9a58-4f04f78034e2-utilities\") pod \"redhat-operators-jl45c\" (UID: \"ff29d64c-c8c9-4cb7-9a58-4f04f78034e2\") " pod="openshift-marketplace/redhat-operators-jl45c" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.561833 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff29d64c-c8c9-4cb7-9a58-4f04f78034e2-catalog-content\") pod \"redhat-operators-jl45c\" (UID: \"ff29d64c-c8c9-4cb7-9a58-4f04f78034e2\") " pod="openshift-marketplace/redhat-operators-jl45c" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.583623 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbcgm\" (UniqueName: \"kubernetes.io/projected/ff29d64c-c8c9-4cb7-9a58-4f04f78034e2-kube-api-access-qbcgm\") pod \"redhat-operators-jl45c\" (UID: \"ff29d64c-c8c9-4cb7-9a58-4f04f78034e2\") " pod="openshift-marketplace/redhat-operators-jl45c" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.727679 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jl45c" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.793251 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wbqf6"] Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.810574 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5c9xs"] Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.834077 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tj8vd" event={"ID":"afc0321c-4482-40a4-92e3-67f5f78f5c6e","Type":"ContainerStarted","Data":"fd0379d318a11896fef2247ef3cf71788c3eae7d610bf703d3fcdbfca59e15b8"} Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.857432 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tj8vd" podStartSLOduration=3.316681221 podStartE2EDuration="8.857407532s" podCreationTimestamp="2026-01-21 18:01:31 +0000 UTC" firstStartedPulling="2026-01-21 18:01:33.673510475 +0000 UTC m=+328.404021360" lastFinishedPulling="2026-01-21 18:01:39.214236786 +0000 UTC m=+333.944747671" observedRunningTime="2026-01-21 18:01:39.852232915 +0000 UTC m=+334.582743820" watchObservedRunningTime="2026-01-21 18:01:39.857407532 +0000 UTC m=+334.587918417" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.861740 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvw6j" event={"ID":"e892666d-d0e9-4169-98d7-70687ac29c1b","Type":"ContainerStarted","Data":"5cca86aee859ae857082332a0432d1a07e803b58da997623d5771df054c4188b"} Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.874083 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pfvs" event={"ID":"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd","Type":"ContainerStarted","Data":"08eadee2b87a03f92f78f5ddf1f293085e3cbb7401c20a8b01b423021c463bac"} Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.893375 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-59dvj" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.902474 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-59dvj" Jan 21 18:01:39 crc kubenswrapper[4808]: I0121 18:01:39.963146 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hzsq2" Jan 21 18:01:40 crc kubenswrapper[4808]: I0121 18:01:40.248176 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jl45c"] Jan 21 18:01:40 crc kubenswrapper[4808]: I0121 18:01:40.881299 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jl45c" event={"ID":"ff29d64c-c8c9-4cb7-9a58-4f04f78034e2","Type":"ContainerStarted","Data":"7b36f9e0cf6b865abfc0978665fc7a3d58a58c051fc00b80308f0dcef1f19b16"} Jan 21 18:01:40 crc kubenswrapper[4808]: I0121 18:01:40.883092 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5c9xs" event={"ID":"21ebff69-0bfe-427d-8330-4e454159dd01","Type":"ContainerStarted","Data":"165fbd2262e741d57cd375773c251eb17b3fa4b02ee5a6658ab73a6334f6108e"} Jan 21 18:01:40 crc kubenswrapper[4808]: I0121 18:01:40.883113 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5c9xs" event={"ID":"21ebff69-0bfe-427d-8330-4e454159dd01","Type":"ContainerStarted","Data":"9040772e5d46bfead7cad29d4879c89e8266e7687fc02b14c492e602792ebf88"} Jan 21 18:01:40 crc kubenswrapper[4808]: I0121 18:01:40.886859 4808 generic.go:334] "Generic (PLEG): container finished" podID="e892666d-d0e9-4169-98d7-70687ac29c1b" containerID="5cca86aee859ae857082332a0432d1a07e803b58da997623d5771df054c4188b" exitCode=0 Jan 21 18:01:40 crc kubenswrapper[4808]: I0121 18:01:40.886931 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvw6j" event={"ID":"e892666d-d0e9-4169-98d7-70687ac29c1b","Type":"ContainerDied","Data":"5cca86aee859ae857082332a0432d1a07e803b58da997623d5771df054c4188b"} Jan 21 18:01:40 crc kubenswrapper[4808]: I0121 18:01:40.891876 4808 generic.go:334] "Generic (PLEG): container finished" podID="1df03daa-d900-4a6d-9776-26f3b05843ce" containerID="4d2be47b39e77a53c6b82949dfa1368760aa914be579e7af148f4593039c2521" exitCode=0 Jan 21 18:01:40 crc kubenswrapper[4808]: I0121 18:01:40.891982 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cbwpb" event={"ID":"1df03daa-d900-4a6d-9776-26f3b05843ce","Type":"ContainerDied","Data":"4d2be47b39e77a53c6b82949dfa1368760aa914be579e7af148f4593039c2521"} Jan 21 18:01:40 crc kubenswrapper[4808]: I0121 18:01:40.899882 4808 generic.go:334] "Generic (PLEG): container finished" podID="9d1c03b0-e10f-47ae-841c-153e4ae0f9cd" containerID="08eadee2b87a03f92f78f5ddf1f293085e3cbb7401c20a8b01b423021c463bac" exitCode=0 Jan 21 18:01:40 crc kubenswrapper[4808]: I0121 18:01:40.899957 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pfvs" event={"ID":"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd","Type":"ContainerDied","Data":"08eadee2b87a03f92f78f5ddf1f293085e3cbb7401c20a8b01b423021c463bac"} Jan 21 18:01:40 crc kubenswrapper[4808]: I0121 18:01:40.904881 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wbqf6" event={"ID":"37705e9c-abf0-42ac-8fc3-03e0b2fa732a","Type":"ContainerStarted","Data":"03aa2b47e0486458355f3c30a167831090b8d72eee394f472cdf2e96ba237ae9"} Jan 21 18:01:40 crc kubenswrapper[4808]: I0121 18:01:40.904919 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wbqf6" event={"ID":"37705e9c-abf0-42ac-8fc3-03e0b2fa732a","Type":"ContainerStarted","Data":"4c2687382daabb8f12e3dcb3b77240b7f6356fbca17a665f14d36e37b9c6df8a"} Jan 21 18:01:40 crc kubenswrapper[4808]: I0121 18:01:40.955481 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-59dvj" podUID="bc01e22f-4eb3-46d9-8bba-faffce8f294e" containerName="registry-server" probeResult="failure" output=< Jan 21 18:01:40 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Jan 21 18:01:40 crc kubenswrapper[4808]: > Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.162297 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kd6dv"] Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.164117 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kd6dv" Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.176704 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kd6dv"] Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.294853 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dde2d0c2-70a4-4d0a-bc7a-29fd72c50095-utilities\") pod \"certified-operators-kd6dv\" (UID: \"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095\") " pod="openshift-marketplace/certified-operators-kd6dv" Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.294900 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dde2d0c2-70a4-4d0a-bc7a-29fd72c50095-catalog-content\") pod \"certified-operators-kd6dv\" (UID: \"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095\") " pod="openshift-marketplace/certified-operators-kd6dv" Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.294960 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mbgq\" (UniqueName: \"kubernetes.io/projected/dde2d0c2-70a4-4d0a-bc7a-29fd72c50095-kube-api-access-7mbgq\") pod \"certified-operators-kd6dv\" (UID: \"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095\") " pod="openshift-marketplace/certified-operators-kd6dv" Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.396770 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mbgq\" (UniqueName: \"kubernetes.io/projected/dde2d0c2-70a4-4d0a-bc7a-29fd72c50095-kube-api-access-7mbgq\") pod \"certified-operators-kd6dv\" (UID: \"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095\") " pod="openshift-marketplace/certified-operators-kd6dv" Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.396914 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dde2d0c2-70a4-4d0a-bc7a-29fd72c50095-utilities\") pod \"certified-operators-kd6dv\" (UID: \"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095\") " pod="openshift-marketplace/certified-operators-kd6dv" Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.396939 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dde2d0c2-70a4-4d0a-bc7a-29fd72c50095-catalog-content\") pod \"certified-operators-kd6dv\" (UID: \"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095\") " pod="openshift-marketplace/certified-operators-kd6dv" Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.397597 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dde2d0c2-70a4-4d0a-bc7a-29fd72c50095-catalog-content\") pod \"certified-operators-kd6dv\" (UID: \"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095\") " pod="openshift-marketplace/certified-operators-kd6dv" Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.398465 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dde2d0c2-70a4-4d0a-bc7a-29fd72c50095-utilities\") pod \"certified-operators-kd6dv\" (UID: \"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095\") " pod="openshift-marketplace/certified-operators-kd6dv" Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.443490 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mbgq\" (UniqueName: \"kubernetes.io/projected/dde2d0c2-70a4-4d0a-bc7a-29fd72c50095-kube-api-access-7mbgq\") pod \"certified-operators-kd6dv\" (UID: \"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095\") " pod="openshift-marketplace/certified-operators-kd6dv" Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.481016 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kd6dv" Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.702973 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xstv7" Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.703021 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xstv7" Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.745723 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xstv7" Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.914494 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pfvs" event={"ID":"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd","Type":"ContainerStarted","Data":"7d1abbc03aa42016ed07b245e1a1ec9e616b4f93dca34fc1da743e210928b23b"} Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.916467 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jl45c" event={"ID":"ff29d64c-c8c9-4cb7-9a58-4f04f78034e2","Type":"ContainerStarted","Data":"92655a705642d3e25a126d87948e676a1b6d2117d9f6948bd0ebbd6700f6ad31"} Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.920575 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hvw6j" event={"ID":"e892666d-d0e9-4169-98d7-70687ac29c1b","Type":"ContainerStarted","Data":"df47edd828baae8832ca1328860004a1264fe46cbcd81b0a3b5cf682a0a0d130"} Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.947284 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4pfvs" podStartSLOduration=5.333672762 podStartE2EDuration="8.947261339s" podCreationTimestamp="2026-01-21 18:01:33 +0000 UTC" firstStartedPulling="2026-01-21 18:01:37.732041599 +0000 UTC m=+332.462552484" lastFinishedPulling="2026-01-21 18:01:41.345630176 +0000 UTC m=+336.076141061" observedRunningTime="2026-01-21 18:01:41.942468952 +0000 UTC m=+336.672979847" watchObservedRunningTime="2026-01-21 18:01:41.947261339 +0000 UTC m=+336.677772244" Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.968975 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2rh88"] Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.970490 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2rh88" Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.980109 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hvw6j" podStartSLOduration=4.395365108 podStartE2EDuration="7.98008044s" podCreationTimestamp="2026-01-21 18:01:34 +0000 UTC" firstStartedPulling="2026-01-21 18:01:37.739593743 +0000 UTC m=+332.470104628" lastFinishedPulling="2026-01-21 18:01:41.324309075 +0000 UTC m=+336.054819960" observedRunningTime="2026-01-21 18:01:41.968115218 +0000 UTC m=+336.698626113" watchObservedRunningTime="2026-01-21 18:01:41.98008044 +0000 UTC m=+336.710591325" Jan 21 18:01:41 crc kubenswrapper[4808]: I0121 18:01:41.987433 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xstv7" Jan 21 18:01:42 crc kubenswrapper[4808]: I0121 18:01:42.034846 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2rh88"] Jan 21 18:01:42 crc kubenswrapper[4808]: I0121 18:01:42.049513 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kd6dv"] Jan 21 18:01:42 crc kubenswrapper[4808]: I0121 18:01:42.108884 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/541d9739-547c-4697-9658-b29bc90e20c6-catalog-content\") pod \"redhat-operators-2rh88\" (UID: \"541d9739-547c-4697-9658-b29bc90e20c6\") " pod="openshift-marketplace/redhat-operators-2rh88" Jan 21 18:01:42 crc kubenswrapper[4808]: I0121 18:01:42.108963 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/541d9739-547c-4697-9658-b29bc90e20c6-utilities\") pod \"redhat-operators-2rh88\" (UID: \"541d9739-547c-4697-9658-b29bc90e20c6\") " pod="openshift-marketplace/redhat-operators-2rh88" Jan 21 18:01:42 crc kubenswrapper[4808]: I0121 18:01:42.109025 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwqwd\" (UniqueName: \"kubernetes.io/projected/541d9739-547c-4697-9658-b29bc90e20c6-kube-api-access-jwqwd\") pod \"redhat-operators-2rh88\" (UID: \"541d9739-547c-4697-9658-b29bc90e20c6\") " pod="openshift-marketplace/redhat-operators-2rh88" Jan 21 18:01:42 crc kubenswrapper[4808]: I0121 18:01:42.210403 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/541d9739-547c-4697-9658-b29bc90e20c6-utilities\") pod \"redhat-operators-2rh88\" (UID: \"541d9739-547c-4697-9658-b29bc90e20c6\") " pod="openshift-marketplace/redhat-operators-2rh88" Jan 21 18:01:42 crc kubenswrapper[4808]: I0121 18:01:42.210676 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwqwd\" (UniqueName: \"kubernetes.io/projected/541d9739-547c-4697-9658-b29bc90e20c6-kube-api-access-jwqwd\") pod \"redhat-operators-2rh88\" (UID: \"541d9739-547c-4697-9658-b29bc90e20c6\") " pod="openshift-marketplace/redhat-operators-2rh88" Jan 21 18:01:42 crc kubenswrapper[4808]: I0121 18:01:42.210901 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/541d9739-547c-4697-9658-b29bc90e20c6-catalog-content\") pod \"redhat-operators-2rh88\" (UID: \"541d9739-547c-4697-9658-b29bc90e20c6\") " pod="openshift-marketplace/redhat-operators-2rh88" Jan 21 18:01:42 crc kubenswrapper[4808]: I0121 18:01:42.210953 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/541d9739-547c-4697-9658-b29bc90e20c6-utilities\") pod \"redhat-operators-2rh88\" (UID: \"541d9739-547c-4697-9658-b29bc90e20c6\") " pod="openshift-marketplace/redhat-operators-2rh88" Jan 21 18:01:42 crc kubenswrapper[4808]: I0121 18:01:42.211407 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/541d9739-547c-4697-9658-b29bc90e20c6-catalog-content\") pod \"redhat-operators-2rh88\" (UID: \"541d9739-547c-4697-9658-b29bc90e20c6\") " pod="openshift-marketplace/redhat-operators-2rh88" Jan 21 18:01:42 crc kubenswrapper[4808]: I0121 18:01:42.236750 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwqwd\" (UniqueName: \"kubernetes.io/projected/541d9739-547c-4697-9658-b29bc90e20c6-kube-api-access-jwqwd\") pod \"redhat-operators-2rh88\" (UID: \"541d9739-547c-4697-9658-b29bc90e20c6\") " pod="openshift-marketplace/redhat-operators-2rh88" Jan 21 18:01:42 crc kubenswrapper[4808]: I0121 18:01:42.290556 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2rh88" Jan 21 18:01:42 crc kubenswrapper[4808]: I0121 18:01:42.295209 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tj8vd" Jan 21 18:01:42 crc kubenswrapper[4808]: I0121 18:01:42.296639 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tj8vd" Jan 21 18:01:42 crc kubenswrapper[4808]: I0121 18:01:42.730160 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2rh88"] Jan 21 18:01:42 crc kubenswrapper[4808]: I0121 18:01:42.925667 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rh88" event={"ID":"541d9739-547c-4697-9658-b29bc90e20c6","Type":"ContainerStarted","Data":"5ed25f42106d3a32ca76d1c9a333dcf2f0ab47c62c1cae23f6041aba1b8a913d"} Jan 21 18:01:42 crc kubenswrapper[4808]: I0121 18:01:42.926604 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kd6dv" event={"ID":"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095","Type":"ContainerStarted","Data":"be07c45071d3263924afee213f2475c067cbb0b5ca5af893e19b86b606764c78"} Jan 21 18:01:43 crc kubenswrapper[4808]: I0121 18:01:43.332211 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tj8vd" podUID="afc0321c-4482-40a4-92e3-67f5f78f5c6e" containerName="registry-server" probeResult="failure" output=< Jan 21 18:01:43 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Jan 21 18:01:43 crc kubenswrapper[4808]: > Jan 21 18:01:43 crc kubenswrapper[4808]: I0121 18:01:43.765314 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nqzlp"] Jan 21 18:01:43 crc kubenswrapper[4808]: I0121 18:01:43.766547 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nqzlp" Jan 21 18:01:43 crc kubenswrapper[4808]: I0121 18:01:43.781177 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nqzlp"] Jan 21 18:01:43 crc kubenswrapper[4808]: I0121 18:01:43.833082 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5322a4c-c457-48c5-9c8d-6ce242b4b87b-catalog-content\") pod \"certified-operators-nqzlp\" (UID: \"c5322a4c-c457-48c5-9c8d-6ce242b4b87b\") " pod="openshift-marketplace/certified-operators-nqzlp" Jan 21 18:01:43 crc kubenswrapper[4808]: I0121 18:01:43.833215 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55m49\" (UniqueName: \"kubernetes.io/projected/c5322a4c-c457-48c5-9c8d-6ce242b4b87b-kube-api-access-55m49\") pod \"certified-operators-nqzlp\" (UID: \"c5322a4c-c457-48c5-9c8d-6ce242b4b87b\") " pod="openshift-marketplace/certified-operators-nqzlp" Jan 21 18:01:43 crc kubenswrapper[4808]: I0121 18:01:43.833260 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5322a4c-c457-48c5-9c8d-6ce242b4b87b-utilities\") pod \"certified-operators-nqzlp\" (UID: \"c5322a4c-c457-48c5-9c8d-6ce242b4b87b\") " pod="openshift-marketplace/certified-operators-nqzlp" Jan 21 18:01:43 crc kubenswrapper[4808]: I0121 18:01:43.934086 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5322a4c-c457-48c5-9c8d-6ce242b4b87b-catalog-content\") pod \"certified-operators-nqzlp\" (UID: \"c5322a4c-c457-48c5-9c8d-6ce242b4b87b\") " pod="openshift-marketplace/certified-operators-nqzlp" Jan 21 18:01:43 crc kubenswrapper[4808]: I0121 18:01:43.934218 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55m49\" (UniqueName: \"kubernetes.io/projected/c5322a4c-c457-48c5-9c8d-6ce242b4b87b-kube-api-access-55m49\") pod \"certified-operators-nqzlp\" (UID: \"c5322a4c-c457-48c5-9c8d-6ce242b4b87b\") " pod="openshift-marketplace/certified-operators-nqzlp" Jan 21 18:01:43 crc kubenswrapper[4808]: I0121 18:01:43.934273 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5322a4c-c457-48c5-9c8d-6ce242b4b87b-utilities\") pod \"certified-operators-nqzlp\" (UID: \"c5322a4c-c457-48c5-9c8d-6ce242b4b87b\") " pod="openshift-marketplace/certified-operators-nqzlp" Jan 21 18:01:43 crc kubenswrapper[4808]: I0121 18:01:43.934692 4808 generic.go:334] "Generic (PLEG): container finished" podID="21ebff69-0bfe-427d-8330-4e454159dd01" containerID="165fbd2262e741d57cd375773c251eb17b3fa4b02ee5a6658ab73a6334f6108e" exitCode=0 Jan 21 18:01:43 crc kubenswrapper[4808]: I0121 18:01:43.934831 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5c9xs" event={"ID":"21ebff69-0bfe-427d-8330-4e454159dd01","Type":"ContainerDied","Data":"165fbd2262e741d57cd375773c251eb17b3fa4b02ee5a6658ab73a6334f6108e"} Jan 21 18:01:43 crc kubenswrapper[4808]: I0121 18:01:43.935954 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5322a4c-c457-48c5-9c8d-6ce242b4b87b-utilities\") pod \"certified-operators-nqzlp\" (UID: \"c5322a4c-c457-48c5-9c8d-6ce242b4b87b\") " pod="openshift-marketplace/certified-operators-nqzlp" Jan 21 18:01:43 crc kubenswrapper[4808]: I0121 18:01:43.936048 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5322a4c-c457-48c5-9c8d-6ce242b4b87b-catalog-content\") pod \"certified-operators-nqzlp\" (UID: \"c5322a4c-c457-48c5-9c8d-6ce242b4b87b\") " pod="openshift-marketplace/certified-operators-nqzlp" Jan 21 18:01:43 crc kubenswrapper[4808]: I0121 18:01:43.936517 4808 generic.go:334] "Generic (PLEG): container finished" podID="37705e9c-abf0-42ac-8fc3-03e0b2fa732a" containerID="03aa2b47e0486458355f3c30a167831090b8d72eee394f472cdf2e96ba237ae9" exitCode=0 Jan 21 18:01:43 crc kubenswrapper[4808]: I0121 18:01:43.936556 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wbqf6" event={"ID":"37705e9c-abf0-42ac-8fc3-03e0b2fa732a","Type":"ContainerDied","Data":"03aa2b47e0486458355f3c30a167831090b8d72eee394f472cdf2e96ba237ae9"} Jan 21 18:01:43 crc kubenswrapper[4808]: I0121 18:01:43.957856 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55m49\" (UniqueName: \"kubernetes.io/projected/c5322a4c-c457-48c5-9c8d-6ce242b4b87b-kube-api-access-55m49\") pod \"certified-operators-nqzlp\" (UID: \"c5322a4c-c457-48c5-9c8d-6ce242b4b87b\") " pod="openshift-marketplace/certified-operators-nqzlp" Jan 21 18:01:44 crc kubenswrapper[4808]: I0121 18:01:44.082781 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nqzlp" Jan 21 18:01:44 crc kubenswrapper[4808]: I0121 18:01:44.307356 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4pfvs" Jan 21 18:01:44 crc kubenswrapper[4808]: I0121 18:01:44.307768 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4pfvs" Jan 21 18:01:44 crc kubenswrapper[4808]: I0121 18:01:44.371696 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4pfvs" Jan 21 18:01:44 crc kubenswrapper[4808]: I0121 18:01:44.627609 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nqzlp"] Jan 21 18:01:44 crc kubenswrapper[4808]: W0121 18:01:44.636085 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5322a4c_c457_48c5_9c8d_6ce242b4b87b.slice/crio-b514c2248d867d2c92ce129f2344ff5d506b91de25738057f1f196306a558a92 WatchSource:0}: Error finding container b514c2248d867d2c92ce129f2344ff5d506b91de25738057f1f196306a558a92: Status 404 returned error can't find the container with id b514c2248d867d2c92ce129f2344ff5d506b91de25738057f1f196306a558a92 Jan 21 18:01:44 crc kubenswrapper[4808]: I0121 18:01:44.895222 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hvw6j" Jan 21 18:01:44 crc kubenswrapper[4808]: I0121 18:01:44.895345 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hvw6j" Jan 21 18:01:44 crc kubenswrapper[4808]: I0121 18:01:44.943503 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqzlp" event={"ID":"c5322a4c-c457-48c5-9c8d-6ce242b4b87b","Type":"ContainerStarted","Data":"b514c2248d867d2c92ce129f2344ff5d506b91de25738057f1f196306a558a92"} Jan 21 18:01:44 crc kubenswrapper[4808]: I0121 18:01:44.945196 4808 generic.go:334] "Generic (PLEG): container finished" podID="dde2d0c2-70a4-4d0a-bc7a-29fd72c50095" containerID="3d6c6e1a6563688f3d99ece628abdd6b6a6294db7fd5393c3979c7540d78fcc6" exitCode=0 Jan 21 18:01:44 crc kubenswrapper[4808]: I0121 18:01:44.945283 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kd6dv" event={"ID":"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095","Type":"ContainerDied","Data":"3d6c6e1a6563688f3d99ece628abdd6b6a6294db7fd5393c3979c7540d78fcc6"} Jan 21 18:01:44 crc kubenswrapper[4808]: I0121 18:01:44.949713 4808 generic.go:334] "Generic (PLEG): container finished" podID="541d9739-547c-4697-9658-b29bc90e20c6" containerID="8e3fc8d411d6cfe17e7fa8dadc52e2c8ee82229cff41aff3fd119bd634e8aceb" exitCode=0 Jan 21 18:01:44 crc kubenswrapper[4808]: I0121 18:01:44.950488 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rh88" event={"ID":"541d9739-547c-4697-9658-b29bc90e20c6","Type":"ContainerDied","Data":"8e3fc8d411d6cfe17e7fa8dadc52e2c8ee82229cff41aff3fd119bd634e8aceb"} Jan 21 18:01:45 crc kubenswrapper[4808]: I0121 18:01:45.931440 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hvw6j" podUID="e892666d-d0e9-4169-98d7-70687ac29c1b" containerName="registry-server" probeResult="failure" output=< Jan 21 18:01:45 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Jan 21 18:01:45 crc kubenswrapper[4808]: > Jan 21 18:01:46 crc kubenswrapper[4808]: I0121 18:01:46.362677 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bwpzc"] Jan 21 18:01:46 crc kubenswrapper[4808]: I0121 18:01:46.364213 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bwpzc" Jan 21 18:01:46 crc kubenswrapper[4808]: I0121 18:01:46.376557 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bwpzc"] Jan 21 18:01:46 crc kubenswrapper[4808]: I0121 18:01:46.477722 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e41b3557-3ab9-43d1-b895-53834748a458-utilities\") pod \"redhat-operators-bwpzc\" (UID: \"e41b3557-3ab9-43d1-b895-53834748a458\") " pod="openshift-marketplace/redhat-operators-bwpzc" Jan 21 18:01:46 crc kubenswrapper[4808]: I0121 18:01:46.478022 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp8d8\" (UniqueName: \"kubernetes.io/projected/e41b3557-3ab9-43d1-b895-53834748a458-kube-api-access-tp8d8\") pod \"redhat-operators-bwpzc\" (UID: \"e41b3557-3ab9-43d1-b895-53834748a458\") " pod="openshift-marketplace/redhat-operators-bwpzc" Jan 21 18:01:46 crc kubenswrapper[4808]: I0121 18:01:46.478191 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e41b3557-3ab9-43d1-b895-53834748a458-catalog-content\") pod \"redhat-operators-bwpzc\" (UID: \"e41b3557-3ab9-43d1-b895-53834748a458\") " pod="openshift-marketplace/redhat-operators-bwpzc" Jan 21 18:01:46 crc kubenswrapper[4808]: I0121 18:01:46.579489 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e41b3557-3ab9-43d1-b895-53834748a458-catalog-content\") pod \"redhat-operators-bwpzc\" (UID: \"e41b3557-3ab9-43d1-b895-53834748a458\") " pod="openshift-marketplace/redhat-operators-bwpzc" Jan 21 18:01:46 crc kubenswrapper[4808]: I0121 18:01:46.579612 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e41b3557-3ab9-43d1-b895-53834748a458-utilities\") pod \"redhat-operators-bwpzc\" (UID: \"e41b3557-3ab9-43d1-b895-53834748a458\") " pod="openshift-marketplace/redhat-operators-bwpzc" Jan 21 18:01:46 crc kubenswrapper[4808]: I0121 18:01:46.579678 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp8d8\" (UniqueName: \"kubernetes.io/projected/e41b3557-3ab9-43d1-b895-53834748a458-kube-api-access-tp8d8\") pod \"redhat-operators-bwpzc\" (UID: \"e41b3557-3ab9-43d1-b895-53834748a458\") " pod="openshift-marketplace/redhat-operators-bwpzc" Jan 21 18:01:46 crc kubenswrapper[4808]: I0121 18:01:46.580541 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e41b3557-3ab9-43d1-b895-53834748a458-catalog-content\") pod \"redhat-operators-bwpzc\" (UID: \"e41b3557-3ab9-43d1-b895-53834748a458\") " pod="openshift-marketplace/redhat-operators-bwpzc" Jan 21 18:01:46 crc kubenswrapper[4808]: I0121 18:01:46.580570 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e41b3557-3ab9-43d1-b895-53834748a458-utilities\") pod \"redhat-operators-bwpzc\" (UID: \"e41b3557-3ab9-43d1-b895-53834748a458\") " pod="openshift-marketplace/redhat-operators-bwpzc" Jan 21 18:01:46 crc kubenswrapper[4808]: I0121 18:01:46.605335 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp8d8\" (UniqueName: \"kubernetes.io/projected/e41b3557-3ab9-43d1-b895-53834748a458-kube-api-access-tp8d8\") pod \"redhat-operators-bwpzc\" (UID: \"e41b3557-3ab9-43d1-b895-53834748a458\") " pod="openshift-marketplace/redhat-operators-bwpzc" Jan 21 18:01:46 crc kubenswrapper[4808]: I0121 18:01:46.681580 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bwpzc" Jan 21 18:01:46 crc kubenswrapper[4808]: I0121 18:01:46.962177 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqzlp" event={"ID":"c5322a4c-c457-48c5-9c8d-6ce242b4b87b","Type":"ContainerStarted","Data":"aa5669676685ea416374092bb35b3e74e779c7133d7b0a8a58c00ce8fbe57f0b"} Jan 21 18:01:46 crc kubenswrapper[4808]: I0121 18:01:46.963936 4808 generic.go:334] "Generic (PLEG): container finished" podID="ff29d64c-c8c9-4cb7-9a58-4f04f78034e2" containerID="92655a705642d3e25a126d87948e676a1b6d2117d9f6948bd0ebbd6700f6ad31" exitCode=0 Jan 21 18:01:46 crc kubenswrapper[4808]: I0121 18:01:46.963978 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jl45c" event={"ID":"ff29d64c-c8c9-4cb7-9a58-4f04f78034e2","Type":"ContainerDied","Data":"92655a705642d3e25a126d87948e676a1b6d2117d9f6948bd0ebbd6700f6ad31"} Jan 21 18:01:47 crc kubenswrapper[4808]: I0121 18:01:47.336971 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bwpzc"] Jan 21 18:01:47 crc kubenswrapper[4808]: W0121 18:01:47.345778 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode41b3557_3ab9_43d1_b895_53834748a458.slice/crio-1ec825028295a686a9ecbd7339b1d7493cb82632a4a8b6ee54eac2f6544039b3 WatchSource:0}: Error finding container 1ec825028295a686a9ecbd7339b1d7493cb82632a4a8b6ee54eac2f6544039b3: Status 404 returned error can't find the container with id 1ec825028295a686a9ecbd7339b1d7493cb82632a4a8b6ee54eac2f6544039b3 Jan 21 18:01:47 crc kubenswrapper[4808]: I0121 18:01:47.543640 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-54shl" Jan 21 18:01:47 crc kubenswrapper[4808]: I0121 18:01:47.590360 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-54shl" Jan 21 18:01:47 crc kubenswrapper[4808]: I0121 18:01:47.975974 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwpzc" event={"ID":"e41b3557-3ab9-43d1-b895-53834748a458","Type":"ContainerStarted","Data":"db1c9c38a72520e1222471c9e369fa109b57dafe121e77c147e2864e336fda7b"} Jan 21 18:01:47 crc kubenswrapper[4808]: I0121 18:01:47.976074 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwpzc" event={"ID":"e41b3557-3ab9-43d1-b895-53834748a458","Type":"ContainerStarted","Data":"1ec825028295a686a9ecbd7339b1d7493cb82632a4a8b6ee54eac2f6544039b3"} Jan 21 18:01:47 crc kubenswrapper[4808]: I0121 18:01:47.979489 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cbwpb" event={"ID":"1df03daa-d900-4a6d-9776-26f3b05843ce","Type":"ContainerStarted","Data":"3845f363524ae982813cf7f417027b238906b51381442ab15ce25a9390788ab0"} Jan 21 18:01:48 crc kubenswrapper[4808]: I0121 18:01:48.986696 4808 generic.go:334] "Generic (PLEG): container finished" podID="1df03daa-d900-4a6d-9776-26f3b05843ce" containerID="3845f363524ae982813cf7f417027b238906b51381442ab15ce25a9390788ab0" exitCode=0 Jan 21 18:01:48 crc kubenswrapper[4808]: I0121 18:01:48.986747 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cbwpb" event={"ID":"1df03daa-d900-4a6d-9776-26f3b05843ce","Type":"ContainerDied","Data":"3845f363524ae982813cf7f417027b238906b51381442ab15ce25a9390788ab0"} Jan 21 18:01:48 crc kubenswrapper[4808]: I0121 18:01:48.988163 4808 generic.go:334] "Generic (PLEG): container finished" podID="c5322a4c-c457-48c5-9c8d-6ce242b4b87b" containerID="aa5669676685ea416374092bb35b3e74e779c7133d7b0a8a58c00ce8fbe57f0b" exitCode=0 Jan 21 18:01:48 crc kubenswrapper[4808]: I0121 18:01:48.988259 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqzlp" event={"ID":"c5322a4c-c457-48c5-9c8d-6ce242b4b87b","Type":"ContainerDied","Data":"aa5669676685ea416374092bb35b3e74e779c7133d7b0a8a58c00ce8fbe57f0b"} Jan 21 18:01:49 crc kubenswrapper[4808]: I0121 18:01:49.948393 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-59dvj" Jan 21 18:01:50 crc kubenswrapper[4808]: I0121 18:01:50.005229 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-59dvj" Jan 21 18:01:51 crc kubenswrapper[4808]: I0121 18:01:51.004680 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5c9xs" event={"ID":"21ebff69-0bfe-427d-8330-4e454159dd01","Type":"ContainerStarted","Data":"a6605e09313eafff3dbcd22afed6ee2336c16ae8ec7853efc4be5ee96817229c"} Jan 21 18:01:52 crc kubenswrapper[4808]: I0121 18:01:52.012903 4808 generic.go:334] "Generic (PLEG): container finished" podID="21ebff69-0bfe-427d-8330-4e454159dd01" containerID="a6605e09313eafff3dbcd22afed6ee2336c16ae8ec7853efc4be5ee96817229c" exitCode=0 Jan 21 18:01:52 crc kubenswrapper[4808]: I0121 18:01:52.012960 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5c9xs" event={"ID":"21ebff69-0bfe-427d-8330-4e454159dd01","Type":"ContainerDied","Data":"a6605e09313eafff3dbcd22afed6ee2336c16ae8ec7853efc4be5ee96817229c"} Jan 21 18:01:52 crc kubenswrapper[4808]: I0121 18:01:52.014943 4808 generic.go:334] "Generic (PLEG): container finished" podID="e41b3557-3ab9-43d1-b895-53834748a458" containerID="db1c9c38a72520e1222471c9e369fa109b57dafe121e77c147e2864e336fda7b" exitCode=0 Jan 21 18:01:52 crc kubenswrapper[4808]: I0121 18:01:52.015038 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwpzc" event={"ID":"e41b3557-3ab9-43d1-b895-53834748a458","Type":"ContainerDied","Data":"db1c9c38a72520e1222471c9e369fa109b57dafe121e77c147e2864e336fda7b"} Jan 21 18:01:52 crc kubenswrapper[4808]: I0121 18:01:52.017161 4808 generic.go:334] "Generic (PLEG): container finished" podID="dde2d0c2-70a4-4d0a-bc7a-29fd72c50095" containerID="bfdb88542a84b2bb64980b75574f46a75b9be79cd15bf33511410187c75f8198" exitCode=0 Jan 21 18:01:52 crc kubenswrapper[4808]: I0121 18:01:52.017218 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kd6dv" event={"ID":"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095","Type":"ContainerDied","Data":"bfdb88542a84b2bb64980b75574f46a75b9be79cd15bf33511410187c75f8198"} Jan 21 18:01:52 crc kubenswrapper[4808]: I0121 18:01:52.019469 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jl45c" event={"ID":"ff29d64c-c8c9-4cb7-9a58-4f04f78034e2","Type":"ContainerStarted","Data":"8bb81a05894809ed85b04b77b5af6aca5437422e753593a5e4394a499db506c8"} Jan 21 18:01:52 crc kubenswrapper[4808]: I0121 18:01:52.341436 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tj8vd" Jan 21 18:01:52 crc kubenswrapper[4808]: I0121 18:01:52.384667 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tj8vd" Jan 21 18:01:53 crc kubenswrapper[4808]: I0121 18:01:53.026316 4808 generic.go:334] "Generic (PLEG): container finished" podID="ff29d64c-c8c9-4cb7-9a58-4f04f78034e2" containerID="8bb81a05894809ed85b04b77b5af6aca5437422e753593a5e4394a499db506c8" exitCode=0 Jan 21 18:01:53 crc kubenswrapper[4808]: I0121 18:01:53.026386 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jl45c" event={"ID":"ff29d64c-c8c9-4cb7-9a58-4f04f78034e2","Type":"ContainerDied","Data":"8bb81a05894809ed85b04b77b5af6aca5437422e753593a5e4394a499db506c8"} Jan 21 18:01:54 crc kubenswrapper[4808]: I0121 18:01:54.376411 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4pfvs" Jan 21 18:01:54 crc kubenswrapper[4808]: I0121 18:01:54.955055 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hvw6j" Jan 21 18:01:55 crc kubenswrapper[4808]: I0121 18:01:55.041627 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hvw6j" Jan 21 18:01:55 crc kubenswrapper[4808]: I0121 18:01:55.058075 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqzlp" event={"ID":"c5322a4c-c457-48c5-9c8d-6ce242b4b87b","Type":"ContainerStarted","Data":"515f64447fd320d6815e2fc89ec6992b1efcd2fed75bcc9bab0e96bce7064366"} Jan 21 18:01:55 crc kubenswrapper[4808]: I0121 18:01:55.076470 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wbqf6" event={"ID":"37705e9c-abf0-42ac-8fc3-03e0b2fa732a","Type":"ContainerStarted","Data":"df122e4d915da59e56d5c8eaad2ef60ed436d9889d449926ef0b0b3b467a9b56"} Jan 21 18:01:55 crc kubenswrapper[4808]: I0121 18:01:55.079264 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kd6dv" event={"ID":"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095","Type":"ContainerStarted","Data":"86efaa6b013918beeef0b0844278cc5c3993935af964fe0cf6acf0db50a968f8"} Jan 21 18:01:55 crc kubenswrapper[4808]: I0121 18:01:55.115752 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5c9xs" event={"ID":"21ebff69-0bfe-427d-8330-4e454159dd01","Type":"ContainerStarted","Data":"1838d953c319f8dde63af6b7d6e8d967718bfa31750f291021a420a8ced264b8"} Jan 21 18:01:55 crc kubenswrapper[4808]: I0121 18:01:55.126018 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cbwpb" event={"ID":"1df03daa-d900-4a6d-9776-26f3b05843ce","Type":"ContainerStarted","Data":"fa7c5bf063b9350cd7c662bbd4575389c72037d13e2efddf22bc2e186973b0cb"} Jan 21 18:01:55 crc kubenswrapper[4808]: I0121 18:01:55.128377 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwpzc" event={"ID":"e41b3557-3ab9-43d1-b895-53834748a458","Type":"ContainerStarted","Data":"da9f737360517c75c63d9c6b8ab16fb7f92c84fd4a7749e5121ca4981301094b"} Jan 21 18:01:55 crc kubenswrapper[4808]: I0121 18:01:55.136689 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rh88" event={"ID":"541d9739-547c-4697-9658-b29bc90e20c6","Type":"ContainerStarted","Data":"e98ff7514f0de42e4aa5992a38b1258d195c5e8698121614d3a37dff62d4aa3f"} Jan 21 18:01:55 crc kubenswrapper[4808]: I0121 18:01:55.142879 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kd6dv" podStartSLOduration=4.309571659 podStartE2EDuration="14.142865748s" podCreationTimestamp="2026-01-21 18:01:41 +0000 UTC" firstStartedPulling="2026-01-21 18:01:44.949055605 +0000 UTC m=+339.679566490" lastFinishedPulling="2026-01-21 18:01:54.782349694 +0000 UTC m=+349.512860579" observedRunningTime="2026-01-21 18:01:55.137173239 +0000 UTC m=+349.867684124" watchObservedRunningTime="2026-01-21 18:01:55.142865748 +0000 UTC m=+349.873376633" Jan 21 18:01:55 crc kubenswrapper[4808]: I0121 18:01:55.186389 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5c9xs" podStartSLOduration=6.238748949 podStartE2EDuration="17.186371561s" podCreationTimestamp="2026-01-21 18:01:38 +0000 UTC" firstStartedPulling="2026-01-21 18:01:43.938137387 +0000 UTC m=+338.668648282" lastFinishedPulling="2026-01-21 18:01:54.885759999 +0000 UTC m=+349.616270894" observedRunningTime="2026-01-21 18:01:55.183631214 +0000 UTC m=+349.914142109" watchObservedRunningTime="2026-01-21 18:01:55.186371561 +0000 UTC m=+349.916882446" Jan 21 18:01:55 crc kubenswrapper[4808]: I0121 18:01:55.240734 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cbwpb" podStartSLOduration=6.062267098 podStartE2EDuration="19.240704078s" podCreationTimestamp="2026-01-21 18:01:36 +0000 UTC" firstStartedPulling="2026-01-21 18:01:40.894866688 +0000 UTC m=+335.625377573" lastFinishedPulling="2026-01-21 18:01:54.073303668 +0000 UTC m=+348.803814553" observedRunningTime="2026-01-21 18:01:55.237597032 +0000 UTC m=+349.968107927" watchObservedRunningTime="2026-01-21 18:01:55.240704078 +0000 UTC m=+349.971214963" Jan 21 18:01:56 crc kubenswrapper[4808]: I0121 18:01:56.143314 4808 generic.go:334] "Generic (PLEG): container finished" podID="541d9739-547c-4697-9658-b29bc90e20c6" containerID="e98ff7514f0de42e4aa5992a38b1258d195c5e8698121614d3a37dff62d4aa3f" exitCode=0 Jan 21 18:01:56 crc kubenswrapper[4808]: I0121 18:01:56.143391 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rh88" event={"ID":"541d9739-547c-4697-9658-b29bc90e20c6","Type":"ContainerDied","Data":"e98ff7514f0de42e4aa5992a38b1258d195c5e8698121614d3a37dff62d4aa3f"} Jan 21 18:01:56 crc kubenswrapper[4808]: I0121 18:01:56.145181 4808 generic.go:334] "Generic (PLEG): container finished" podID="c5322a4c-c457-48c5-9c8d-6ce242b4b87b" containerID="515f64447fd320d6815e2fc89ec6992b1efcd2fed75bcc9bab0e96bce7064366" exitCode=0 Jan 21 18:01:56 crc kubenswrapper[4808]: I0121 18:01:56.145292 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqzlp" event={"ID":"c5322a4c-c457-48c5-9c8d-6ce242b4b87b","Type":"ContainerDied","Data":"515f64447fd320d6815e2fc89ec6992b1efcd2fed75bcc9bab0e96bce7064366"} Jan 21 18:01:56 crc kubenswrapper[4808]: I0121 18:01:56.147763 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jl45c" event={"ID":"ff29d64c-c8c9-4cb7-9a58-4f04f78034e2","Type":"ContainerStarted","Data":"37970e07835436d1823133157468f689a2cac59d3071b6b878dc68fe4e170b22"} Jan 21 18:01:56 crc kubenswrapper[4808]: I0121 18:01:56.188362 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jl45c" podStartSLOduration=9.205375828 podStartE2EDuration="17.18834228s" podCreationTimestamp="2026-01-21 18:01:39 +0000 UTC" firstStartedPulling="2026-01-21 18:01:46.96554207 +0000 UTC m=+341.696052955" lastFinishedPulling="2026-01-21 18:01:54.948508502 +0000 UTC m=+349.679019407" observedRunningTime="2026-01-21 18:01:56.186136756 +0000 UTC m=+350.916647661" watchObservedRunningTime="2026-01-21 18:01:56.18834228 +0000 UTC m=+350.918853165" Jan 21 18:01:57 crc kubenswrapper[4808]: I0121 18:01:56.698618 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cbwpb" Jan 21 18:01:57 crc kubenswrapper[4808]: I0121 18:01:56.698972 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cbwpb" Jan 21 18:01:57 crc kubenswrapper[4808]: I0121 18:01:57.156954 4808 generic.go:334] "Generic (PLEG): container finished" podID="e41b3557-3ab9-43d1-b895-53834748a458" containerID="da9f737360517c75c63d9c6b8ab16fb7f92c84fd4a7749e5121ca4981301094b" exitCode=0 Jan 21 18:01:57 crc kubenswrapper[4808]: I0121 18:01:57.157026 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwpzc" event={"ID":"e41b3557-3ab9-43d1-b895-53834748a458","Type":"ContainerDied","Data":"da9f737360517c75c63d9c6b8ab16fb7f92c84fd4a7749e5121ca4981301094b"} Jan 21 18:01:57 crc kubenswrapper[4808]: I0121 18:01:57.160885 4808 generic.go:334] "Generic (PLEG): container finished" podID="37705e9c-abf0-42ac-8fc3-03e0b2fa732a" containerID="df122e4d915da59e56d5c8eaad2ef60ed436d9889d449926ef0b0b3b467a9b56" exitCode=0 Jan 21 18:01:57 crc kubenswrapper[4808]: I0121 18:01:57.160931 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wbqf6" event={"ID":"37705e9c-abf0-42ac-8fc3-03e0b2fa732a","Type":"ContainerDied","Data":"df122e4d915da59e56d5c8eaad2ef60ed436d9889d449926ef0b0b3b467a9b56"} Jan 21 18:01:57 crc kubenswrapper[4808]: I0121 18:01:57.186955 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cbwpb" Jan 21 18:01:59 crc kubenswrapper[4808]: I0121 18:01:59.175812 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqzlp" event={"ID":"c5322a4c-c457-48c5-9c8d-6ce242b4b87b","Type":"ContainerStarted","Data":"48bce33ad57d52f0f446058cfd1ba8b2673b597e34eed6ca33bb88dda0af606c"} Jan 21 18:01:59 crc kubenswrapper[4808]: I0121 18:01:59.177730 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wbqf6" event={"ID":"37705e9c-abf0-42ac-8fc3-03e0b2fa732a","Type":"ContainerStarted","Data":"58706841cb670a8e96ca0d588e9cb91d2e51752568ce2f8648f533f1b59e2fe4"} Jan 21 18:01:59 crc kubenswrapper[4808]: I0121 18:01:59.179443 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwpzc" event={"ID":"e41b3557-3ab9-43d1-b895-53834748a458","Type":"ContainerStarted","Data":"645c3581a0a6ce502e3d5080f290d10dc2d2066310e73664d945fe43d1711bf5"} Jan 21 18:01:59 crc kubenswrapper[4808]: I0121 18:01:59.181017 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rh88" event={"ID":"541d9739-547c-4697-9658-b29bc90e20c6","Type":"ContainerStarted","Data":"5531a39bd12a11987c66c43f6f6bab8d1ad397c2ed135606def3dc7b99ea8574"} Jan 21 18:01:59 crc kubenswrapper[4808]: I0121 18:01:59.200403 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nqzlp" podStartSLOduration=8.696077874 podStartE2EDuration="16.200380027s" podCreationTimestamp="2026-01-21 18:01:43 +0000 UTC" firstStartedPulling="2026-01-21 18:01:50.508175694 +0000 UTC m=+345.238686589" lastFinishedPulling="2026-01-21 18:01:58.012477867 +0000 UTC m=+352.742988742" observedRunningTime="2026-01-21 18:01:59.199128866 +0000 UTC m=+353.929639751" watchObservedRunningTime="2026-01-21 18:01:59.200380027 +0000 UTC m=+353.930890912" Jan 21 18:01:59 crc kubenswrapper[4808]: I0121 18:01:59.224534 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wbqf6" podStartSLOduration=8.896280766 podStartE2EDuration="23.224513756s" podCreationTimestamp="2026-01-21 18:01:36 +0000 UTC" firstStartedPulling="2026-01-21 18:01:43.938135657 +0000 UTC m=+338.668646562" lastFinishedPulling="2026-01-21 18:01:58.266368667 +0000 UTC m=+352.996879552" observedRunningTime="2026-01-21 18:01:59.219386001 +0000 UTC m=+353.949896906" watchObservedRunningTime="2026-01-21 18:01:59.224513756 +0000 UTC m=+353.955024641" Jan 21 18:01:59 crc kubenswrapper[4808]: I0121 18:01:59.240869 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2rh88" podStartSLOduration=6.089247162 podStartE2EDuration="18.240850825s" podCreationTimestamp="2026-01-21 18:01:41 +0000 UTC" firstStartedPulling="2026-01-21 18:01:45.957157815 +0000 UTC m=+340.687668690" lastFinishedPulling="2026-01-21 18:01:58.108761468 +0000 UTC m=+352.839272353" observedRunningTime="2026-01-21 18:01:59.239136853 +0000 UTC m=+353.969647738" watchObservedRunningTime="2026-01-21 18:01:59.240850825 +0000 UTC m=+353.971361700" Jan 21 18:01:59 crc kubenswrapper[4808]: I0121 18:01:59.258071 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bwpzc" podStartSLOduration=7.164016993 podStartE2EDuration="13.258053425s" podCreationTimestamp="2026-01-21 18:01:46 +0000 UTC" firstStartedPulling="2026-01-21 18:01:52.182093423 +0000 UTC m=+346.912604308" lastFinishedPulling="2026-01-21 18:01:58.276129855 +0000 UTC m=+353.006640740" observedRunningTime="2026-01-21 18:01:59.255413471 +0000 UTC m=+353.985924356" watchObservedRunningTime="2026-01-21 18:01:59.258053425 +0000 UTC m=+353.988564310" Jan 21 18:01:59 crc kubenswrapper[4808]: I0121 18:01:59.275416 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5c9xs" Jan 21 18:01:59 crc kubenswrapper[4808]: I0121 18:01:59.275455 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5c9xs" Jan 21 18:01:59 crc kubenswrapper[4808]: I0121 18:01:59.316564 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5c9xs" Jan 21 18:01:59 crc kubenswrapper[4808]: I0121 18:01:59.762411 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jl45c" Jan 21 18:01:59 crc kubenswrapper[4808]: I0121 18:01:59.763589 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jl45c" Jan 21 18:02:00 crc kubenswrapper[4808]: I0121 18:02:00.233008 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5c9xs" Jan 21 18:02:00 crc kubenswrapper[4808]: I0121 18:02:00.803952 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jl45c" podUID="ff29d64c-c8c9-4cb7-9a58-4f04f78034e2" containerName="registry-server" probeResult="failure" output=< Jan 21 18:02:00 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Jan 21 18:02:00 crc kubenswrapper[4808]: > Jan 21 18:02:01 crc kubenswrapper[4808]: I0121 18:02:01.481997 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kd6dv" Jan 21 18:02:01 crc kubenswrapper[4808]: I0121 18:02:01.482344 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kd6dv" Jan 21 18:02:01 crc kubenswrapper[4808]: I0121 18:02:01.535025 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kd6dv" Jan 21 18:02:02 crc kubenswrapper[4808]: I0121 18:02:02.240218 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kd6dv" Jan 21 18:02:02 crc kubenswrapper[4808]: I0121 18:02:02.291553 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2rh88" Jan 21 18:02:02 crc kubenswrapper[4808]: I0121 18:02:02.292221 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2rh88" Jan 21 18:02:03 crc kubenswrapper[4808]: I0121 18:02:03.344011 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2rh88" podUID="541d9739-547c-4697-9658-b29bc90e20c6" containerName="registry-server" probeResult="failure" output=< Jan 21 18:02:03 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Jan 21 18:02:03 crc kubenswrapper[4808]: > Jan 21 18:02:04 crc kubenswrapper[4808]: I0121 18:02:04.083337 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nqzlp" Jan 21 18:02:04 crc kubenswrapper[4808]: I0121 18:02:04.083394 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nqzlp" Jan 21 18:02:04 crc kubenswrapper[4808]: I0121 18:02:04.124077 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nqzlp" Jan 21 18:02:04 crc kubenswrapper[4808]: I0121 18:02:04.245730 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nqzlp" Jan 21 18:02:06 crc kubenswrapper[4808]: I0121 18:02:06.681800 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bwpzc" Jan 21 18:02:06 crc kubenswrapper[4808]: I0121 18:02:06.682953 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bwpzc" Jan 21 18:02:06 crc kubenswrapper[4808]: I0121 18:02:06.740717 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bwpzc" Jan 21 18:02:06 crc kubenswrapper[4808]: I0121 18:02:06.779167 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cbwpb" Jan 21 18:02:07 crc kubenswrapper[4808]: I0121 18:02:07.282389 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bwpzc" Jan 21 18:02:07 crc kubenswrapper[4808]: I0121 18:02:07.289970 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wbqf6" Jan 21 18:02:07 crc kubenswrapper[4808]: I0121 18:02:07.290132 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wbqf6" Jan 21 18:02:07 crc kubenswrapper[4808]: I0121 18:02:07.350863 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wbqf6" Jan 21 18:02:08 crc kubenswrapper[4808]: I0121 18:02:08.282339 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wbqf6" Jan 21 18:02:09 crc kubenswrapper[4808]: I0121 18:02:09.766182 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jl45c" Jan 21 18:02:09 crc kubenswrapper[4808]: I0121 18:02:09.813503 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jl45c" Jan 21 18:02:12 crc kubenswrapper[4808]: I0121 18:02:12.332921 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2rh88" Jan 21 18:02:12 crc kubenswrapper[4808]: I0121 18:02:12.373686 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2rh88" Jan 21 18:02:29 crc kubenswrapper[4808]: I0121 18:02:29.599921 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:02:29 crc kubenswrapper[4808]: I0121 18:02:29.600746 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:02:59 crc kubenswrapper[4808]: I0121 18:02:59.599695 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:02:59 crc kubenswrapper[4808]: I0121 18:02:59.600531 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:03:29 crc kubenswrapper[4808]: I0121 18:03:29.600613 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:03:29 crc kubenswrapper[4808]: I0121 18:03:29.601270 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:03:29 crc kubenswrapper[4808]: I0121 18:03:29.601322 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 18:03:29 crc kubenswrapper[4808]: I0121 18:03:29.601901 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b1f94a9dbae93a06cda5ec2fa475694e62e99f13f0c15187ffea4d991182c486"} pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 18:03:29 crc kubenswrapper[4808]: I0121 18:03:29.601947 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" containerID="cri-o://b1f94a9dbae93a06cda5ec2fa475694e62e99f13f0c15187ffea4d991182c486" gracePeriod=600 Jan 21 18:03:29 crc kubenswrapper[4808]: I0121 18:03:29.825417 4808 generic.go:334] "Generic (PLEG): container finished" podID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerID="b1f94a9dbae93a06cda5ec2fa475694e62e99f13f0c15187ffea4d991182c486" exitCode=0 Jan 21 18:03:29 crc kubenswrapper[4808]: I0121 18:03:29.825484 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerDied","Data":"b1f94a9dbae93a06cda5ec2fa475694e62e99f13f0c15187ffea4d991182c486"} Jan 21 18:03:29 crc kubenswrapper[4808]: I0121 18:03:29.825948 4808 scope.go:117] "RemoveContainer" containerID="d06b6608b8a0caa31c6595a6e9bdc35d34a6a99c29e0613acfbc18d00c3c49cb" Jan 21 18:03:30 crc kubenswrapper[4808]: I0121 18:03:30.836537 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerStarted","Data":"00f0bd827addaebc04ffc5544dc603b3fe979bfa986819c93eca5d866ec75282"} Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.029130 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fc2fk"] Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.030564 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="nbdb" containerID="cri-o://1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f" gracePeriod=30 Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.030610 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="northd" containerID="cri-o://25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd" gracePeriod=30 Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.030703 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="kube-rbac-proxy-node" containerID="cri-o://b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739" gracePeriod=30 Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.030994 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovn-acl-logging" containerID="cri-o://c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53" gracePeriod=30 Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.031055 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="sbdb" containerID="cri-o://23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae" gracePeriod=30 Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.030573 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53" gracePeriod=30 Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.030540 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovn-controller" containerID="cri-o://5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053" gracePeriod=30 Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.095329 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovnkube-controller" containerID="cri-o://ac3a70f942488a03cabf1cce2d89342bfda3f785a417e6ec0246ff1a1f715c71" gracePeriod=30 Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.508280 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovnkube-controller/3.log" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.511091 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovn-acl-logging/0.log" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.512062 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovn-controller/0.log" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.512794 4808 generic.go:334] "Generic (PLEG): container finished" podID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerID="ac3a70f942488a03cabf1cce2d89342bfda3f785a417e6ec0246ff1a1f715c71" exitCode=0 Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.512823 4808 generic.go:334] "Generic (PLEG): container finished" podID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerID="23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae" exitCode=0 Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.512837 4808 generic.go:334] "Generic (PLEG): container finished" podID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerID="1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f" exitCode=0 Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.512848 4808 generic.go:334] "Generic (PLEG): container finished" podID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerID="25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd" exitCode=0 Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.512842 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerDied","Data":"ac3a70f942488a03cabf1cce2d89342bfda3f785a417e6ec0246ff1a1f715c71"} Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.512884 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerDied","Data":"23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae"} Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.512897 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerDied","Data":"1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f"} Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.512911 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerDied","Data":"25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd"} Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.512919 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerDied","Data":"6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53"} Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.512856 4808 generic.go:334] "Generic (PLEG): container finished" podID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerID="6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53" exitCode=0 Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.512922 4808 scope.go:117] "RemoveContainer" containerID="02b5440e1f9434c7998a819e1e62e4d10ad6887724bac037b63d65847c039f48" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.512941 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerDied","Data":"b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739"} Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.512935 4808 generic.go:334] "Generic (PLEG): container finished" podID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerID="b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739" exitCode=0 Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.513080 4808 generic.go:334] "Generic (PLEG): container finished" podID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerID="c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53" exitCode=143 Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.513125 4808 generic.go:334] "Generic (PLEG): container finished" podID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerID="5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053" exitCode=143 Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.513144 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerDied","Data":"c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53"} Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.513177 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerDied","Data":"5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053"} Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.516193 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8fp9_538d1d44-6c67-450a-809c-d9008128ec9f/kube-multus/2.log" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.517032 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8fp9_538d1d44-6c67-450a-809c-d9008128ec9f/kube-multus/1.log" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.517124 4808 generic.go:334] "Generic (PLEG): container finished" podID="538d1d44-6c67-450a-809c-d9008128ec9f" containerID="e9f2431a8d4f410cded26b98a893f5432e81290e6947a1807e97cb72619b8c01" exitCode=2 Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.517177 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8fp9" event={"ID":"538d1d44-6c67-450a-809c-d9008128ec9f","Type":"ContainerDied","Data":"e9f2431a8d4f410cded26b98a893f5432e81290e6947a1807e97cb72619b8c01"} Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.518071 4808 scope.go:117] "RemoveContainer" containerID="e9f2431a8d4f410cded26b98a893f5432e81290e6947a1807e97cb72619b8c01" Jan 21 18:05:03 crc kubenswrapper[4808]: E0121 18:05:03.518597 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-z8fp9_openshift-multus(538d1d44-6c67-450a-809c-d9008128ec9f)\"" pod="openshift-multus/multus-z8fp9" podUID="538d1d44-6c67-450a-809c-d9008128ec9f" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.566896 4808 scope.go:117] "RemoveContainer" containerID="f9074967ffe383fc367aaa834087e74f523d9563ecae59d79810bb389f05fbdc" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.741747 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovn-acl-logging/0.log" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.742345 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovn-controller/0.log" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.742978 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.817758 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-knbws"] Jan 21 18:05:03 crc kubenswrapper[4808]: E0121 18:05:03.818086 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="northd" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818102 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="northd" Jan 21 18:05:03 crc kubenswrapper[4808]: E0121 18:05:03.818123 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="kube-rbac-proxy-ovn-metrics" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818132 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="kube-rbac-proxy-ovn-metrics" Jan 21 18:05:03 crc kubenswrapper[4808]: E0121 18:05:03.818141 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovnkube-controller" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818154 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovnkube-controller" Jan 21 18:05:03 crc kubenswrapper[4808]: E0121 18:05:03.818163 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="sbdb" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818171 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="sbdb" Jan 21 18:05:03 crc kubenswrapper[4808]: E0121 18:05:03.818181 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovnkube-controller" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818188 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovnkube-controller" Jan 21 18:05:03 crc kubenswrapper[4808]: E0121 18:05:03.818196 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovnkube-controller" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818204 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovnkube-controller" Jan 21 18:05:03 crc kubenswrapper[4808]: E0121 18:05:03.818217 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovn-acl-logging" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818224 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovn-acl-logging" Jan 21 18:05:03 crc kubenswrapper[4808]: E0121 18:05:03.818235 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="nbdb" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818278 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="nbdb" Jan 21 18:05:03 crc kubenswrapper[4808]: E0121 18:05:03.818289 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovn-controller" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818297 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovn-controller" Jan 21 18:05:03 crc kubenswrapper[4808]: E0121 18:05:03.818310 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="kubecfg-setup" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818320 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="kubecfg-setup" Jan 21 18:05:03 crc kubenswrapper[4808]: E0121 18:05:03.818337 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="kube-rbac-proxy-node" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818346 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="kube-rbac-proxy-node" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818480 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovnkube-controller" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818493 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovnkube-controller" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818503 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="nbdb" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818514 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="kube-rbac-proxy-node" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818522 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="sbdb" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818534 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovn-controller" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818545 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovnkube-controller" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818554 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="kube-rbac-proxy-ovn-metrics" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818567 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="northd" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818579 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovnkube-controller" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818588 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovn-acl-logging" Jan 21 18:05:03 crc kubenswrapper[4808]: E0121 18:05:03.818720 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovnkube-controller" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818729 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovnkube-controller" Jan 21 18:05:03 crc kubenswrapper[4808]: E0121 18:05:03.818740 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovnkube-controller" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818750 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovnkube-controller" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.818885 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" containerName="ovnkube-controller" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.821448 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.911742 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bb13a20f-f4ab-412c-8165-5b93edc79628-ovnkube-config\") pod \"bb13a20f-f4ab-412c-8165-5b93edc79628\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.911791 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bb13a20f-f4ab-412c-8165-5b93edc79628-env-overrides\") pod \"bb13a20f-f4ab-412c-8165-5b93edc79628\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.911814 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-kubelet\") pod \"bb13a20f-f4ab-412c-8165-5b93edc79628\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.911841 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-var-lib-openvswitch\") pod \"bb13a20f-f4ab-412c-8165-5b93edc79628\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.911870 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-cni-bin\") pod \"bb13a20f-f4ab-412c-8165-5b93edc79628\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.911902 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-run-ovn-kubernetes\") pod \"bb13a20f-f4ab-412c-8165-5b93edc79628\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.911923 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bb13a20f-f4ab-412c-8165-5b93edc79628-ovn-node-metrics-cert\") pod \"bb13a20f-f4ab-412c-8165-5b93edc79628\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.911962 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-cni-netd\") pod \"bb13a20f-f4ab-412c-8165-5b93edc79628\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.911996 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-etc-openvswitch\") pod \"bb13a20f-f4ab-412c-8165-5b93edc79628\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.912020 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-slash\") pod \"bb13a20f-f4ab-412c-8165-5b93edc79628\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.912039 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-node-log\") pod \"bb13a20f-f4ab-412c-8165-5b93edc79628\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.912055 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-run-netns\") pod \"bb13a20f-f4ab-412c-8165-5b93edc79628\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.912085 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-run-openvswitch\") pod \"bb13a20f-f4ab-412c-8165-5b93edc79628\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.912108 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-run-systemd\") pod \"bb13a20f-f4ab-412c-8165-5b93edc79628\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.912131 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bb13a20f-f4ab-412c-8165-5b93edc79628-ovnkube-script-lib\") pod \"bb13a20f-f4ab-412c-8165-5b93edc79628\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.912149 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-log-socket\") pod \"bb13a20f-f4ab-412c-8165-5b93edc79628\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.912169 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-systemd-units\") pod \"bb13a20f-f4ab-412c-8165-5b93edc79628\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.912186 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwp7h\" (UniqueName: \"kubernetes.io/projected/bb13a20f-f4ab-412c-8165-5b93edc79628-kube-api-access-zwp7h\") pod \"bb13a20f-f4ab-412c-8165-5b93edc79628\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.912203 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-var-lib-cni-networks-ovn-kubernetes\") pod \"bb13a20f-f4ab-412c-8165-5b93edc79628\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.912228 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-run-ovn\") pod \"bb13a20f-f4ab-412c-8165-5b93edc79628\" (UID: \"bb13a20f-f4ab-412c-8165-5b93edc79628\") " Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.912446 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-slash" (OuterVolumeSpecName: "host-slash") pod "bb13a20f-f4ab-412c-8165-5b93edc79628" (UID: "bb13a20f-f4ab-412c-8165-5b93edc79628"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.912495 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "bb13a20f-f4ab-412c-8165-5b93edc79628" (UID: "bb13a20f-f4ab-412c-8165-5b93edc79628"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.912524 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-log-socket" (OuterVolumeSpecName: "log-socket") pod "bb13a20f-f4ab-412c-8165-5b93edc79628" (UID: "bb13a20f-f4ab-412c-8165-5b93edc79628"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.912531 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "bb13a20f-f4ab-412c-8165-5b93edc79628" (UID: "bb13a20f-f4ab-412c-8165-5b93edc79628"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.912546 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "bb13a20f-f4ab-412c-8165-5b93edc79628" (UID: "bb13a20f-f4ab-412c-8165-5b93edc79628"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.912576 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-node-log" (OuterVolumeSpecName: "node-log") pod "bb13a20f-f4ab-412c-8165-5b93edc79628" (UID: "bb13a20f-f4ab-412c-8165-5b93edc79628"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.912641 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "bb13a20f-f4ab-412c-8165-5b93edc79628" (UID: "bb13a20f-f4ab-412c-8165-5b93edc79628"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.913043 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb13a20f-f4ab-412c-8165-5b93edc79628-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "bb13a20f-f4ab-412c-8165-5b93edc79628" (UID: "bb13a20f-f4ab-412c-8165-5b93edc79628"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.913062 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb13a20f-f4ab-412c-8165-5b93edc79628-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "bb13a20f-f4ab-412c-8165-5b93edc79628" (UID: "bb13a20f-f4ab-412c-8165-5b93edc79628"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.913080 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "bb13a20f-f4ab-412c-8165-5b93edc79628" (UID: "bb13a20f-f4ab-412c-8165-5b93edc79628"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.913079 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb13a20f-f4ab-412c-8165-5b93edc79628-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "bb13a20f-f4ab-412c-8165-5b93edc79628" (UID: "bb13a20f-f4ab-412c-8165-5b93edc79628"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.913099 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "bb13a20f-f4ab-412c-8165-5b93edc79628" (UID: "bb13a20f-f4ab-412c-8165-5b93edc79628"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.913124 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "bb13a20f-f4ab-412c-8165-5b93edc79628" (UID: "bb13a20f-f4ab-412c-8165-5b93edc79628"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.913155 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "bb13a20f-f4ab-412c-8165-5b93edc79628" (UID: "bb13a20f-f4ab-412c-8165-5b93edc79628"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.913181 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "bb13a20f-f4ab-412c-8165-5b93edc79628" (UID: "bb13a20f-f4ab-412c-8165-5b93edc79628"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.913204 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "bb13a20f-f4ab-412c-8165-5b93edc79628" (UID: "bb13a20f-f4ab-412c-8165-5b93edc79628"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.913226 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "bb13a20f-f4ab-412c-8165-5b93edc79628" (UID: "bb13a20f-f4ab-412c-8165-5b93edc79628"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.920173 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb13a20f-f4ab-412c-8165-5b93edc79628-kube-api-access-zwp7h" (OuterVolumeSpecName: "kube-api-access-zwp7h") pod "bb13a20f-f4ab-412c-8165-5b93edc79628" (UID: "bb13a20f-f4ab-412c-8165-5b93edc79628"). InnerVolumeSpecName "kube-api-access-zwp7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.920390 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb13a20f-f4ab-412c-8165-5b93edc79628-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "bb13a20f-f4ab-412c-8165-5b93edc79628" (UID: "bb13a20f-f4ab-412c-8165-5b93edc79628"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:05:03 crc kubenswrapper[4808]: I0121 18:05:03.932105 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "bb13a20f-f4ab-412c-8165-5b93edc79628" (UID: "bb13a20f-f4ab-412c-8165-5b93edc79628"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.014556 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-log-socket\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.014652 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-cni-bin\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.014850 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/48d54a7c-c3a4-4ff7-a181-9372dcb44188-ovnkube-config\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.014965 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.015021 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-etc-openvswitch\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.015074 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-node-log\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.015236 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-systemd-units\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.015321 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-slash\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.015353 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-var-lib-openvswitch\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.015418 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-run-openvswitch\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.015501 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-cni-netd\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.015567 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-run-systemd\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.015643 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgphc\" (UniqueName: \"kubernetes.io/projected/48d54a7c-c3a4-4ff7-a181-9372dcb44188-kube-api-access-xgphc\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.015699 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-run-ovn\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.015723 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-kubelet\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.015747 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-run-netns\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.015782 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-run-ovn-kubernetes\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.015868 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/48d54a7c-c3a4-4ff7-a181-9372dcb44188-ovnkube-script-lib\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016020 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/48d54a7c-c3a4-4ff7-a181-9372dcb44188-ovn-node-metrics-cert\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016064 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/48d54a7c-c3a4-4ff7-a181-9372dcb44188-env-overrides\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016183 4808 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-kubelet\") on node \"crc\" DevicePath \"\"" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016210 4808 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016227 4808 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-cni-bin\") on node \"crc\" DevicePath \"\"" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016268 4808 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016287 4808 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bb13a20f-f4ab-412c-8165-5b93edc79628-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016299 4808 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-cni-netd\") on node \"crc\" DevicePath \"\"" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016310 4808 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016321 4808 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-slash\") on node \"crc\" DevicePath \"\"" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016331 4808 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-node-log\") on node \"crc\" DevicePath \"\"" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016344 4808 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-run-netns\") on node \"crc\" DevicePath \"\"" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016358 4808 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-run-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016368 4808 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-run-systemd\") on node \"crc\" DevicePath \"\"" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016379 4808 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bb13a20f-f4ab-412c-8165-5b93edc79628-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016391 4808 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-log-socket\") on node \"crc\" DevicePath \"\"" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016401 4808 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-systemd-units\") on node \"crc\" DevicePath \"\"" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016411 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwp7h\" (UniqueName: \"kubernetes.io/projected/bb13a20f-f4ab-412c-8165-5b93edc79628-kube-api-access-zwp7h\") on node \"crc\" DevicePath \"\"" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016422 4808 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016432 4808 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb13a20f-f4ab-412c-8165-5b93edc79628-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016444 4808 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bb13a20f-f4ab-412c-8165-5b93edc79628-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.016454 4808 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bb13a20f-f4ab-412c-8165-5b93edc79628-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.117477 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-cni-netd\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.117547 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-run-systemd\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.117581 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgphc\" (UniqueName: \"kubernetes.io/projected/48d54a7c-c3a4-4ff7-a181-9372dcb44188-kube-api-access-xgphc\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.117609 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-run-ovn\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.117625 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-run-systemd\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.117625 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-cni-netd\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.117688 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-kubelet\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.117717 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-run-ovn\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.117635 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-kubelet\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.117777 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-run-netns\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.117803 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-run-ovn-kubernetes\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.117833 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/48d54a7c-c3a4-4ff7-a181-9372dcb44188-ovnkube-script-lib\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.117842 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-run-netns\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.117877 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-run-ovn-kubernetes\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.117935 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/48d54a7c-c3a4-4ff7-a181-9372dcb44188-ovn-node-metrics-cert\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.117967 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/48d54a7c-c3a4-4ff7-a181-9372dcb44188-env-overrides\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.118009 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-log-socket\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.118054 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-cni-bin\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.118095 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/48d54a7c-c3a4-4ff7-a181-9372dcb44188-ovnkube-config\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.118162 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.118206 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-etc-openvswitch\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.118284 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-node-log\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.118383 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-systemd-units\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.118425 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-var-lib-openvswitch\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.118444 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-slash\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.118485 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-run-openvswitch\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.118585 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/48d54a7c-c3a4-4ff7-a181-9372dcb44188-ovnkube-script-lib\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.118601 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-run-openvswitch\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.118653 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-etc-openvswitch\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.118660 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-cni-bin\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.118725 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-log-socket\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.118757 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-systemd-units\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.118766 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.118805 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-node-log\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.118811 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-host-slash\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.118884 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48d54a7c-c3a4-4ff7-a181-9372dcb44188-var-lib-openvswitch\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.119064 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/48d54a7c-c3a4-4ff7-a181-9372dcb44188-env-overrides\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.119274 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/48d54a7c-c3a4-4ff7-a181-9372dcb44188-ovnkube-config\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.121980 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/48d54a7c-c3a4-4ff7-a181-9372dcb44188-ovn-node-metrics-cert\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.135196 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgphc\" (UniqueName: \"kubernetes.io/projected/48d54a7c-c3a4-4ff7-a181-9372dcb44188-kube-api-access-xgphc\") pod \"ovnkube-node-knbws\" (UID: \"48d54a7c-c3a4-4ff7-a181-9372dcb44188\") " pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.137651 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.527529 4808 generic.go:334] "Generic (PLEG): container finished" podID="48d54a7c-c3a4-4ff7-a181-9372dcb44188" containerID="6752255168e7b150b650cf8fa4c375b77155e5670ff3f0cb3562e3105160659e" exitCode=0 Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.527635 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-knbws" event={"ID":"48d54a7c-c3a4-4ff7-a181-9372dcb44188","Type":"ContainerDied","Data":"6752255168e7b150b650cf8fa4c375b77155e5670ff3f0cb3562e3105160659e"} Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.527716 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-knbws" event={"ID":"48d54a7c-c3a4-4ff7-a181-9372dcb44188","Type":"ContainerStarted","Data":"83efa117ce5efeafd59d65317faa5a92fee164615886190a8ce7c7f0bb8e6125"} Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.530547 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8fp9_538d1d44-6c67-450a-809c-d9008128ec9f/kube-multus/2.log" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.535972 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovn-acl-logging/0.log" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.538283 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc2fk_bb13a20f-f4ab-412c-8165-5b93edc79628/ovn-controller/0.log" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.539072 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" event={"ID":"bb13a20f-f4ab-412c-8165-5b93edc79628","Type":"ContainerDied","Data":"bdceb85fb9701f1d5c24bc358c55d9e3cbc44b1259f1a3f401679dc516c72136"} Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.539123 4808 scope.go:117] "RemoveContainer" containerID="ac3a70f942488a03cabf1cce2d89342bfda3f785a417e6ec0246ff1a1f715c71" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.539365 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fc2fk" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.564551 4808 scope.go:117] "RemoveContainer" containerID="23ea1a18794023ebdc0beaa280e3fd076b63be8652b90d17e32fd3c34a8c85ae" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.609445 4808 scope.go:117] "RemoveContainer" containerID="1cf78ff383b966d5dfa95936b4b77c67b877ce0d2e18ca20f62694487efec17f" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.611357 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fc2fk"] Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.618380 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fc2fk"] Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.634355 4808 scope.go:117] "RemoveContainer" containerID="25b0d89c55a4f4c084aa076cc63a86396ba1a67264f8885cec0ce3ddcb1749cd" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.663963 4808 scope.go:117] "RemoveContainer" containerID="6ba801d05b4627a348e5554f8066ca05a0b25200dab9c10c01d1ea6f65b28c53" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.680770 4808 scope.go:117] "RemoveContainer" containerID="b1c64c43de0c7702ebc5a7c6c938e2b621a3b96104f8c763168b16c38502f739" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.696986 4808 scope.go:117] "RemoveContainer" containerID="c5057670eeeb2f5746e1ca69adb2c4439ee491bc2594e5ed5103739bee934a53" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.727050 4808 scope.go:117] "RemoveContainer" containerID="5edf9d0b17476a60ff667f301dada2577b33f8db25e7c75a49111c9601dd1053" Jan 21 18:05:04 crc kubenswrapper[4808]: I0121 18:05:04.745636 4808 scope.go:117] "RemoveContainer" containerID="d112e0d2fba960c8910b47c05a50420aff059a137a0f64add57655549f411653" Jan 21 18:05:05 crc kubenswrapper[4808]: I0121 18:05:05.542283 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb13a20f-f4ab-412c-8165-5b93edc79628" path="/var/lib/kubelet/pods/bb13a20f-f4ab-412c-8165-5b93edc79628/volumes" Jan 21 18:05:05 crc kubenswrapper[4808]: I0121 18:05:05.564781 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-knbws" event={"ID":"48d54a7c-c3a4-4ff7-a181-9372dcb44188","Type":"ContainerStarted","Data":"1e4b21be1ec2e90d06fc7bc86adf9dbc2b680fd859c8b96795f018a64fac8fb9"} Jan 21 18:05:05 crc kubenswrapper[4808]: I0121 18:05:05.564845 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-knbws" event={"ID":"48d54a7c-c3a4-4ff7-a181-9372dcb44188","Type":"ContainerStarted","Data":"29fabb0883cf52a13a14e2cf8c2eeaa16ca71507469d40d136e22a5a6e373a25"} Jan 21 18:05:05 crc kubenswrapper[4808]: I0121 18:05:05.564861 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-knbws" event={"ID":"48d54a7c-c3a4-4ff7-a181-9372dcb44188","Type":"ContainerStarted","Data":"7bb1199d93274eab4002b285da4e59c469a4efca4cbd5740e9090091556e06fb"} Jan 21 18:05:05 crc kubenswrapper[4808]: I0121 18:05:05.564875 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-knbws" event={"ID":"48d54a7c-c3a4-4ff7-a181-9372dcb44188","Type":"ContainerStarted","Data":"c4da7920e3fa8c3a941daf1a9c2dcbb7d407483bf83627992791997447ef2a76"} Jan 21 18:05:05 crc kubenswrapper[4808]: I0121 18:05:05.564888 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-knbws" event={"ID":"48d54a7c-c3a4-4ff7-a181-9372dcb44188","Type":"ContainerStarted","Data":"99f7abe22f880dbb6f2ebac456da7c6184ceba9b07b6dc186abf24cd5a867c5b"} Jan 21 18:05:05 crc kubenswrapper[4808]: I0121 18:05:05.564902 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-knbws" event={"ID":"48d54a7c-c3a4-4ff7-a181-9372dcb44188","Type":"ContainerStarted","Data":"f0bb79f16f7b9ebf8e70bac657bc7fed4d8f5765a35e7eb736f75310fcdd7938"} Jan 21 18:05:08 crc kubenswrapper[4808]: I0121 18:05:08.589930 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-knbws" event={"ID":"48d54a7c-c3a4-4ff7-a181-9372dcb44188","Type":"ContainerStarted","Data":"eced1a6edb3f1860cc62acdea0722a8a8b1cfc0661ac71edc75d94356c1f335c"} Jan 21 18:05:10 crc kubenswrapper[4808]: I0121 18:05:10.614377 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-knbws" event={"ID":"48d54a7c-c3a4-4ff7-a181-9372dcb44188","Type":"ContainerStarted","Data":"e19a85d4c5d343eabbbc4f5037eb8189436836b1a9cb4dec942935f3e1498274"} Jan 21 18:05:10 crc kubenswrapper[4808]: I0121 18:05:10.616492 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:10 crc kubenswrapper[4808]: I0121 18:05:10.616650 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:10 crc kubenswrapper[4808]: I0121 18:05:10.616717 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:10 crc kubenswrapper[4808]: I0121 18:05:10.645661 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-knbws" podStartSLOduration=7.645638693 podStartE2EDuration="7.645638693s" podCreationTimestamp="2026-01-21 18:05:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 18:05:10.641558537 +0000 UTC m=+545.372069442" watchObservedRunningTime="2026-01-21 18:05:10.645638693 +0000 UTC m=+545.376149588" Jan 21 18:05:10 crc kubenswrapper[4808]: I0121 18:05:10.648731 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:10 crc kubenswrapper[4808]: I0121 18:05:10.651398 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:14 crc kubenswrapper[4808]: I0121 18:05:14.520592 4808 scope.go:117] "RemoveContainer" containerID="e9f2431a8d4f410cded26b98a893f5432e81290e6947a1807e97cb72619b8c01" Jan 21 18:05:14 crc kubenswrapper[4808]: E0121 18:05:14.521395 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-z8fp9_openshift-multus(538d1d44-6c67-450a-809c-d9008128ec9f)\"" pod="openshift-multus/multus-z8fp9" podUID="538d1d44-6c67-450a-809c-d9008128ec9f" Jan 21 18:05:25 crc kubenswrapper[4808]: I0121 18:05:25.526992 4808 scope.go:117] "RemoveContainer" containerID="e9f2431a8d4f410cded26b98a893f5432e81290e6947a1807e97cb72619b8c01" Jan 21 18:05:26 crc kubenswrapper[4808]: I0121 18:05:26.722546 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8fp9_538d1d44-6c67-450a-809c-d9008128ec9f/kube-multus/2.log" Jan 21 18:05:26 crc kubenswrapper[4808]: I0121 18:05:26.722920 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8fp9" event={"ID":"538d1d44-6c67-450a-809c-d9008128ec9f","Type":"ContainerStarted","Data":"78818fda8d8874dede4fbdef5876d431c7856c8a6bed718cd106aec42f301726"} Jan 21 18:05:29 crc kubenswrapper[4808]: I0121 18:05:29.599887 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:05:29 crc kubenswrapper[4808]: I0121 18:05:29.600387 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:05:34 crc kubenswrapper[4808]: I0121 18:05:34.170953 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-knbws" Jan 21 18:05:59 crc kubenswrapper[4808]: I0121 18:05:59.600593 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:05:59 crc kubenswrapper[4808]: I0121 18:05:59.601203 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:06:10 crc kubenswrapper[4808]: I0121 18:06:10.956594 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lx72r"] Jan 21 18:06:10 crc kubenswrapper[4808]: I0121 18:06:10.957530 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lx72r" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" containerName="registry-server" containerID="cri-o://02be607e2450f7f343306b4491c2582fa28e9942e7331e122cf44d05a00d02a1" gracePeriod=30 Jan 21 18:06:11 crc kubenswrapper[4808]: I0121 18:06:11.345964 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lx72r" Jan 21 18:06:11 crc kubenswrapper[4808]: I0121 18:06:11.405653 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhbxc\" (UniqueName: \"kubernetes.io/projected/b05e5192-b448-446a-96ba-3ec53d3644bf-kube-api-access-nhbxc\") pod \"b05e5192-b448-446a-96ba-3ec53d3644bf\" (UID: \"b05e5192-b448-446a-96ba-3ec53d3644bf\") " Jan 21 18:06:11 crc kubenswrapper[4808]: I0121 18:06:11.405755 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b05e5192-b448-446a-96ba-3ec53d3644bf-utilities\") pod \"b05e5192-b448-446a-96ba-3ec53d3644bf\" (UID: \"b05e5192-b448-446a-96ba-3ec53d3644bf\") " Jan 21 18:06:11 crc kubenswrapper[4808]: I0121 18:06:11.405782 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b05e5192-b448-446a-96ba-3ec53d3644bf-catalog-content\") pod \"b05e5192-b448-446a-96ba-3ec53d3644bf\" (UID: \"b05e5192-b448-446a-96ba-3ec53d3644bf\") " Jan 21 18:06:11 crc kubenswrapper[4808]: I0121 18:06:11.406966 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b05e5192-b448-446a-96ba-3ec53d3644bf-utilities" (OuterVolumeSpecName: "utilities") pod "b05e5192-b448-446a-96ba-3ec53d3644bf" (UID: "b05e5192-b448-446a-96ba-3ec53d3644bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:06:11 crc kubenswrapper[4808]: I0121 18:06:11.411444 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b05e5192-b448-446a-96ba-3ec53d3644bf-kube-api-access-nhbxc" (OuterVolumeSpecName: "kube-api-access-nhbxc") pod "b05e5192-b448-446a-96ba-3ec53d3644bf" (UID: "b05e5192-b448-446a-96ba-3ec53d3644bf"). InnerVolumeSpecName "kube-api-access-nhbxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:06:11 crc kubenswrapper[4808]: I0121 18:06:11.427775 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b05e5192-b448-446a-96ba-3ec53d3644bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b05e5192-b448-446a-96ba-3ec53d3644bf" (UID: "b05e5192-b448-446a-96ba-3ec53d3644bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:06:11 crc kubenswrapper[4808]: I0121 18:06:11.508109 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhbxc\" (UniqueName: \"kubernetes.io/projected/b05e5192-b448-446a-96ba-3ec53d3644bf-kube-api-access-nhbxc\") on node \"crc\" DevicePath \"\"" Jan 21 18:06:11 crc kubenswrapper[4808]: I0121 18:06:11.508190 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b05e5192-b448-446a-96ba-3ec53d3644bf-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:06:11 crc kubenswrapper[4808]: I0121 18:06:11.508216 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b05e5192-b448-446a-96ba-3ec53d3644bf-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:06:12 crc kubenswrapper[4808]: I0121 18:06:12.013064 4808 generic.go:334] "Generic (PLEG): container finished" podID="b05e5192-b448-446a-96ba-3ec53d3644bf" containerID="02be607e2450f7f343306b4491c2582fa28e9942e7331e122cf44d05a00d02a1" exitCode=0 Jan 21 18:06:12 crc kubenswrapper[4808]: I0121 18:06:12.013136 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lx72r" event={"ID":"b05e5192-b448-446a-96ba-3ec53d3644bf","Type":"ContainerDied","Data":"02be607e2450f7f343306b4491c2582fa28e9942e7331e122cf44d05a00d02a1"} Jan 21 18:06:12 crc kubenswrapper[4808]: I0121 18:06:12.013521 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lx72r" event={"ID":"b05e5192-b448-446a-96ba-3ec53d3644bf","Type":"ContainerDied","Data":"27d08e2d789e1f390fc3e64743f37ba3962f0a028c36ff21c630239b6dcb13ef"} Jan 21 18:06:12 crc kubenswrapper[4808]: I0121 18:06:12.013552 4808 scope.go:117] "RemoveContainer" containerID="02be607e2450f7f343306b4491c2582fa28e9942e7331e122cf44d05a00d02a1" Jan 21 18:06:12 crc kubenswrapper[4808]: I0121 18:06:12.013178 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lx72r" Jan 21 18:06:12 crc kubenswrapper[4808]: I0121 18:06:12.039928 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lx72r"] Jan 21 18:06:12 crc kubenswrapper[4808]: I0121 18:06:12.045006 4808 scope.go:117] "RemoveContainer" containerID="00f32933828b596a7577e3722d32d68205a1c6b654d7a120a69a25051225ebc1" Jan 21 18:06:12 crc kubenswrapper[4808]: I0121 18:06:12.046038 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lx72r"] Jan 21 18:06:12 crc kubenswrapper[4808]: I0121 18:06:12.062897 4808 scope.go:117] "RemoveContainer" containerID="568239dfedb5eb527404b60490c0913fe635260d28ee528f39bb6e9f8f6f4e05" Jan 21 18:06:12 crc kubenswrapper[4808]: I0121 18:06:12.082466 4808 scope.go:117] "RemoveContainer" containerID="02be607e2450f7f343306b4491c2582fa28e9942e7331e122cf44d05a00d02a1" Jan 21 18:06:12 crc kubenswrapper[4808]: E0121 18:06:12.082992 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02be607e2450f7f343306b4491c2582fa28e9942e7331e122cf44d05a00d02a1\": container with ID starting with 02be607e2450f7f343306b4491c2582fa28e9942e7331e122cf44d05a00d02a1 not found: ID does not exist" containerID="02be607e2450f7f343306b4491c2582fa28e9942e7331e122cf44d05a00d02a1" Jan 21 18:06:12 crc kubenswrapper[4808]: I0121 18:06:12.083065 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02be607e2450f7f343306b4491c2582fa28e9942e7331e122cf44d05a00d02a1"} err="failed to get container status \"02be607e2450f7f343306b4491c2582fa28e9942e7331e122cf44d05a00d02a1\": rpc error: code = NotFound desc = could not find container \"02be607e2450f7f343306b4491c2582fa28e9942e7331e122cf44d05a00d02a1\": container with ID starting with 02be607e2450f7f343306b4491c2582fa28e9942e7331e122cf44d05a00d02a1 not found: ID does not exist" Jan 21 18:06:12 crc kubenswrapper[4808]: I0121 18:06:12.083116 4808 scope.go:117] "RemoveContainer" containerID="00f32933828b596a7577e3722d32d68205a1c6b654d7a120a69a25051225ebc1" Jan 21 18:06:12 crc kubenswrapper[4808]: E0121 18:06:12.083565 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00f32933828b596a7577e3722d32d68205a1c6b654d7a120a69a25051225ebc1\": container with ID starting with 00f32933828b596a7577e3722d32d68205a1c6b654d7a120a69a25051225ebc1 not found: ID does not exist" containerID="00f32933828b596a7577e3722d32d68205a1c6b654d7a120a69a25051225ebc1" Jan 21 18:06:12 crc kubenswrapper[4808]: I0121 18:06:12.083601 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00f32933828b596a7577e3722d32d68205a1c6b654d7a120a69a25051225ebc1"} err="failed to get container status \"00f32933828b596a7577e3722d32d68205a1c6b654d7a120a69a25051225ebc1\": rpc error: code = NotFound desc = could not find container \"00f32933828b596a7577e3722d32d68205a1c6b654d7a120a69a25051225ebc1\": container with ID starting with 00f32933828b596a7577e3722d32d68205a1c6b654d7a120a69a25051225ebc1 not found: ID does not exist" Jan 21 18:06:12 crc kubenswrapper[4808]: I0121 18:06:12.083630 4808 scope.go:117] "RemoveContainer" containerID="568239dfedb5eb527404b60490c0913fe635260d28ee528f39bb6e9f8f6f4e05" Jan 21 18:06:12 crc kubenswrapper[4808]: E0121 18:06:12.084073 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"568239dfedb5eb527404b60490c0913fe635260d28ee528f39bb6e9f8f6f4e05\": container with ID starting with 568239dfedb5eb527404b60490c0913fe635260d28ee528f39bb6e9f8f6f4e05 not found: ID does not exist" containerID="568239dfedb5eb527404b60490c0913fe635260d28ee528f39bb6e9f8f6f4e05" Jan 21 18:06:12 crc kubenswrapper[4808]: I0121 18:06:12.084122 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"568239dfedb5eb527404b60490c0913fe635260d28ee528f39bb6e9f8f6f4e05"} err="failed to get container status \"568239dfedb5eb527404b60490c0913fe635260d28ee528f39bb6e9f8f6f4e05\": rpc error: code = NotFound desc = could not find container \"568239dfedb5eb527404b60490c0913fe635260d28ee528f39bb6e9f8f6f4e05\": container with ID starting with 568239dfedb5eb527404b60490c0913fe635260d28ee528f39bb6e9f8f6f4e05 not found: ID does not exist" Jan 21 18:06:13 crc kubenswrapper[4808]: I0121 18:06:13.528934 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" path="/var/lib/kubelet/pods/b05e5192-b448-446a-96ba-3ec53d3644bf/volumes" Jan 21 18:06:14 crc kubenswrapper[4808]: I0121 18:06:14.821740 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd"] Jan 21 18:06:14 crc kubenswrapper[4808]: E0121 18:06:14.822544 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" containerName="extract-utilities" Jan 21 18:06:14 crc kubenswrapper[4808]: I0121 18:06:14.822560 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" containerName="extract-utilities" Jan 21 18:06:14 crc kubenswrapper[4808]: E0121 18:06:14.822568 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" containerName="extract-content" Jan 21 18:06:14 crc kubenswrapper[4808]: I0121 18:06:14.822575 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" containerName="extract-content" Jan 21 18:06:14 crc kubenswrapper[4808]: E0121 18:06:14.822599 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" containerName="registry-server" Jan 21 18:06:14 crc kubenswrapper[4808]: I0121 18:06:14.822607 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" containerName="registry-server" Jan 21 18:06:14 crc kubenswrapper[4808]: I0121 18:06:14.822738 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="b05e5192-b448-446a-96ba-3ec53d3644bf" containerName="registry-server" Jan 21 18:06:14 crc kubenswrapper[4808]: I0121 18:06:14.823809 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd" Jan 21 18:06:14 crc kubenswrapper[4808]: I0121 18:06:14.827187 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 21 18:06:14 crc kubenswrapper[4808]: I0121 18:06:14.837504 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd"] Jan 21 18:06:14 crc kubenswrapper[4808]: I0121 18:06:14.867033 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zzsh\" (UniqueName: \"kubernetes.io/projected/de6d62bf-8431-46c9-9720-9874644fabd9-kube-api-access-9zzsh\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd\" (UID: \"de6d62bf-8431-46c9-9720-9874644fabd9\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd" Jan 21 18:06:14 crc kubenswrapper[4808]: I0121 18:06:14.867136 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/de6d62bf-8431-46c9-9720-9874644fabd9-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd\" (UID: \"de6d62bf-8431-46c9-9720-9874644fabd9\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd" Jan 21 18:06:14 crc kubenswrapper[4808]: I0121 18:06:14.867165 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/de6d62bf-8431-46c9-9720-9874644fabd9-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd\" (UID: \"de6d62bf-8431-46c9-9720-9874644fabd9\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd" Jan 21 18:06:14 crc kubenswrapper[4808]: I0121 18:06:14.969151 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zzsh\" (UniqueName: \"kubernetes.io/projected/de6d62bf-8431-46c9-9720-9874644fabd9-kube-api-access-9zzsh\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd\" (UID: \"de6d62bf-8431-46c9-9720-9874644fabd9\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd" Jan 21 18:06:14 crc kubenswrapper[4808]: I0121 18:06:14.969268 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/de6d62bf-8431-46c9-9720-9874644fabd9-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd\" (UID: \"de6d62bf-8431-46c9-9720-9874644fabd9\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd" Jan 21 18:06:14 crc kubenswrapper[4808]: I0121 18:06:14.969305 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/de6d62bf-8431-46c9-9720-9874644fabd9-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd\" (UID: \"de6d62bf-8431-46c9-9720-9874644fabd9\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd" Jan 21 18:06:14 crc kubenswrapper[4808]: I0121 18:06:14.969804 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/de6d62bf-8431-46c9-9720-9874644fabd9-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd\" (UID: \"de6d62bf-8431-46c9-9720-9874644fabd9\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd" Jan 21 18:06:14 crc kubenswrapper[4808]: I0121 18:06:14.969993 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/de6d62bf-8431-46c9-9720-9874644fabd9-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd\" (UID: \"de6d62bf-8431-46c9-9720-9874644fabd9\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd" Jan 21 18:06:14 crc kubenswrapper[4808]: I0121 18:06:14.989412 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zzsh\" (UniqueName: \"kubernetes.io/projected/de6d62bf-8431-46c9-9720-9874644fabd9-kube-api-access-9zzsh\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd\" (UID: \"de6d62bf-8431-46c9-9720-9874644fabd9\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd" Jan 21 18:06:15 crc kubenswrapper[4808]: I0121 18:06:15.144963 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd" Jan 21 18:06:15 crc kubenswrapper[4808]: I0121 18:06:15.339155 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd"] Jan 21 18:06:16 crc kubenswrapper[4808]: I0121 18:06:16.048690 4808 generic.go:334] "Generic (PLEG): container finished" podID="de6d62bf-8431-46c9-9720-9874644fabd9" containerID="61cd9561ee8ad9b5df1748950f35b1254ecacfc1ae5b6b06224baacadae6f631" exitCode=0 Jan 21 18:06:16 crc kubenswrapper[4808]: I0121 18:06:16.048776 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd" event={"ID":"de6d62bf-8431-46c9-9720-9874644fabd9","Type":"ContainerDied","Data":"61cd9561ee8ad9b5df1748950f35b1254ecacfc1ae5b6b06224baacadae6f631"} Jan 21 18:06:16 crc kubenswrapper[4808]: I0121 18:06:16.048827 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd" event={"ID":"de6d62bf-8431-46c9-9720-9874644fabd9","Type":"ContainerStarted","Data":"bfba7450e52813c0d5e3a61090f429737844667ea5c5af9eade44ecb89c6b98b"} Jan 21 18:06:16 crc kubenswrapper[4808]: I0121 18:06:16.050520 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 18:06:18 crc kubenswrapper[4808]: I0121 18:06:18.068675 4808 generic.go:334] "Generic (PLEG): container finished" podID="de6d62bf-8431-46c9-9720-9874644fabd9" containerID="fad22b45824af4f56bdd5dae553bbdf69cdf387f9c6a997ed9ddba6689a845fe" exitCode=0 Jan 21 18:06:18 crc kubenswrapper[4808]: I0121 18:06:18.068766 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd" event={"ID":"de6d62bf-8431-46c9-9720-9874644fabd9","Type":"ContainerDied","Data":"fad22b45824af4f56bdd5dae553bbdf69cdf387f9c6a997ed9ddba6689a845fe"} Jan 21 18:06:19 crc kubenswrapper[4808]: I0121 18:06:19.082059 4808 generic.go:334] "Generic (PLEG): container finished" podID="de6d62bf-8431-46c9-9720-9874644fabd9" containerID="3551b86d76d077887555f2c7878e4d6762a8f227666ffaf8c904b7eaae201a8c" exitCode=0 Jan 21 18:06:19 crc kubenswrapper[4808]: I0121 18:06:19.082168 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd" event={"ID":"de6d62bf-8431-46c9-9720-9874644fabd9","Type":"ContainerDied","Data":"3551b86d76d077887555f2c7878e4d6762a8f227666ffaf8c904b7eaae201a8c"} Jan 21 18:06:20 crc kubenswrapper[4808]: I0121 18:06:20.356543 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd" Jan 21 18:06:20 crc kubenswrapper[4808]: I0121 18:06:20.454822 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zzsh\" (UniqueName: \"kubernetes.io/projected/de6d62bf-8431-46c9-9720-9874644fabd9-kube-api-access-9zzsh\") pod \"de6d62bf-8431-46c9-9720-9874644fabd9\" (UID: \"de6d62bf-8431-46c9-9720-9874644fabd9\") " Jan 21 18:06:20 crc kubenswrapper[4808]: I0121 18:06:20.455094 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/de6d62bf-8431-46c9-9720-9874644fabd9-bundle\") pod \"de6d62bf-8431-46c9-9720-9874644fabd9\" (UID: \"de6d62bf-8431-46c9-9720-9874644fabd9\") " Jan 21 18:06:20 crc kubenswrapper[4808]: I0121 18:06:20.455131 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/de6d62bf-8431-46c9-9720-9874644fabd9-util\") pod \"de6d62bf-8431-46c9-9720-9874644fabd9\" (UID: \"de6d62bf-8431-46c9-9720-9874644fabd9\") " Jan 21 18:06:20 crc kubenswrapper[4808]: I0121 18:06:20.458037 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de6d62bf-8431-46c9-9720-9874644fabd9-bundle" (OuterVolumeSpecName: "bundle") pod "de6d62bf-8431-46c9-9720-9874644fabd9" (UID: "de6d62bf-8431-46c9-9720-9874644fabd9"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:06:20 crc kubenswrapper[4808]: I0121 18:06:20.463513 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de6d62bf-8431-46c9-9720-9874644fabd9-kube-api-access-9zzsh" (OuterVolumeSpecName: "kube-api-access-9zzsh") pod "de6d62bf-8431-46c9-9720-9874644fabd9" (UID: "de6d62bf-8431-46c9-9720-9874644fabd9"). InnerVolumeSpecName "kube-api-access-9zzsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:06:20 crc kubenswrapper[4808]: I0121 18:06:20.469273 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de6d62bf-8431-46c9-9720-9874644fabd9-util" (OuterVolumeSpecName: "util") pod "de6d62bf-8431-46c9-9720-9874644fabd9" (UID: "de6d62bf-8431-46c9-9720-9874644fabd9"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:06:20 crc kubenswrapper[4808]: I0121 18:06:20.556184 4808 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/de6d62bf-8431-46c9-9720-9874644fabd9-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 18:06:20 crc kubenswrapper[4808]: I0121 18:06:20.556227 4808 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/de6d62bf-8431-46c9-9720-9874644fabd9-util\") on node \"crc\" DevicePath \"\"" Jan 21 18:06:20 crc kubenswrapper[4808]: I0121 18:06:20.556282 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zzsh\" (UniqueName: \"kubernetes.io/projected/de6d62bf-8431-46c9-9720-9874644fabd9-kube-api-access-9zzsh\") on node \"crc\" DevicePath \"\"" Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.099691 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd" event={"ID":"de6d62bf-8431-46c9-9720-9874644fabd9","Type":"ContainerDied","Data":"bfba7450e52813c0d5e3a61090f429737844667ea5c5af9eade44ecb89c6b98b"} Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.099739 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfba7450e52813c0d5e3a61090f429737844667ea5c5af9eade44ecb89c6b98b" Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.100148 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd" Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.200996 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm"] Jan 21 18:06:21 crc kubenswrapper[4808]: E0121 18:06:21.201271 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de6d62bf-8431-46c9-9720-9874644fabd9" containerName="util" Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.201284 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="de6d62bf-8431-46c9-9720-9874644fabd9" containerName="util" Jan 21 18:06:21 crc kubenswrapper[4808]: E0121 18:06:21.201306 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de6d62bf-8431-46c9-9720-9874644fabd9" containerName="pull" Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.201313 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="de6d62bf-8431-46c9-9720-9874644fabd9" containerName="pull" Jan 21 18:06:21 crc kubenswrapper[4808]: E0121 18:06:21.201322 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de6d62bf-8431-46c9-9720-9874644fabd9" containerName="extract" Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.201328 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="de6d62bf-8431-46c9-9720-9874644fabd9" containerName="extract" Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.201464 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="de6d62bf-8431-46c9-9720-9874644fabd9" containerName="extract" Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.202475 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm" Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.205227 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.214962 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm"] Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.263553 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8935b136-3537-4e51-ae90-9c3729258689-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm\" (UID: \"8935b136-3537-4e51-ae90-9c3729258689\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm" Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.263691 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6fjp\" (UniqueName: \"kubernetes.io/projected/8935b136-3537-4e51-ae90-9c3729258689-kube-api-access-p6fjp\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm\" (UID: \"8935b136-3537-4e51-ae90-9c3729258689\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm" Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.263730 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8935b136-3537-4e51-ae90-9c3729258689-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm\" (UID: \"8935b136-3537-4e51-ae90-9c3729258689\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm" Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.365470 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6fjp\" (UniqueName: \"kubernetes.io/projected/8935b136-3537-4e51-ae90-9c3729258689-kube-api-access-p6fjp\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm\" (UID: \"8935b136-3537-4e51-ae90-9c3729258689\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm" Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.365530 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8935b136-3537-4e51-ae90-9c3729258689-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm\" (UID: \"8935b136-3537-4e51-ae90-9c3729258689\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm" Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.365618 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8935b136-3537-4e51-ae90-9c3729258689-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm\" (UID: \"8935b136-3537-4e51-ae90-9c3729258689\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm" Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.366276 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8935b136-3537-4e51-ae90-9c3729258689-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm\" (UID: \"8935b136-3537-4e51-ae90-9c3729258689\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm" Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.366314 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8935b136-3537-4e51-ae90-9c3729258689-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm\" (UID: \"8935b136-3537-4e51-ae90-9c3729258689\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm" Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.385501 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6fjp\" (UniqueName: \"kubernetes.io/projected/8935b136-3537-4e51-ae90-9c3729258689-kube-api-access-p6fjp\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm\" (UID: \"8935b136-3537-4e51-ae90-9c3729258689\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm" Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.519914 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm" Jan 21 18:06:21 crc kubenswrapper[4808]: I0121 18:06:21.907768 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm"] Jan 21 18:06:22 crc kubenswrapper[4808]: I0121 18:06:22.000607 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9"] Jan 21 18:06:22 crc kubenswrapper[4808]: I0121 18:06:22.001761 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9" Jan 21 18:06:22 crc kubenswrapper[4808]: I0121 18:06:22.014967 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9"] Jan 21 18:06:22 crc kubenswrapper[4808]: I0121 18:06:22.075450 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnshv\" (UniqueName: \"kubernetes.io/projected/61b09e55-6f9f-4dd0-a174-8fcd696335c2-kube-api-access-tnshv\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9\" (UID: \"61b09e55-6f9f-4dd0-a174-8fcd696335c2\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9" Jan 21 18:06:22 crc kubenswrapper[4808]: I0121 18:06:22.075574 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/61b09e55-6f9f-4dd0-a174-8fcd696335c2-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9\" (UID: \"61b09e55-6f9f-4dd0-a174-8fcd696335c2\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9" Jan 21 18:06:22 crc kubenswrapper[4808]: I0121 18:06:22.075667 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/61b09e55-6f9f-4dd0-a174-8fcd696335c2-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9\" (UID: \"61b09e55-6f9f-4dd0-a174-8fcd696335c2\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9" Jan 21 18:06:22 crc kubenswrapper[4808]: I0121 18:06:22.106660 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm" event={"ID":"8935b136-3537-4e51-ae90-9c3729258689","Type":"ContainerStarted","Data":"d4b0faf92f9aa608eb22842b99877514ce45ccf670276a7db21fdf73944afcf4"} Jan 21 18:06:22 crc kubenswrapper[4808]: I0121 18:06:22.106727 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm" event={"ID":"8935b136-3537-4e51-ae90-9c3729258689","Type":"ContainerStarted","Data":"2b50cfb238a2d21aedf9f21c91efcf783990badacc259892f32c009b3115fe8a"} Jan 21 18:06:22 crc kubenswrapper[4808]: I0121 18:06:22.176473 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnshv\" (UniqueName: \"kubernetes.io/projected/61b09e55-6f9f-4dd0-a174-8fcd696335c2-kube-api-access-tnshv\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9\" (UID: \"61b09e55-6f9f-4dd0-a174-8fcd696335c2\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9" Jan 21 18:06:22 crc kubenswrapper[4808]: I0121 18:06:22.176549 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/61b09e55-6f9f-4dd0-a174-8fcd696335c2-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9\" (UID: \"61b09e55-6f9f-4dd0-a174-8fcd696335c2\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9" Jan 21 18:06:22 crc kubenswrapper[4808]: I0121 18:06:22.176637 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/61b09e55-6f9f-4dd0-a174-8fcd696335c2-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9\" (UID: \"61b09e55-6f9f-4dd0-a174-8fcd696335c2\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9" Jan 21 18:06:22 crc kubenswrapper[4808]: I0121 18:06:22.177033 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/61b09e55-6f9f-4dd0-a174-8fcd696335c2-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9\" (UID: \"61b09e55-6f9f-4dd0-a174-8fcd696335c2\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9" Jan 21 18:06:22 crc kubenswrapper[4808]: I0121 18:06:22.177134 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/61b09e55-6f9f-4dd0-a174-8fcd696335c2-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9\" (UID: \"61b09e55-6f9f-4dd0-a174-8fcd696335c2\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9" Jan 21 18:06:22 crc kubenswrapper[4808]: I0121 18:06:22.193433 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnshv\" (UniqueName: \"kubernetes.io/projected/61b09e55-6f9f-4dd0-a174-8fcd696335c2-kube-api-access-tnshv\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9\" (UID: \"61b09e55-6f9f-4dd0-a174-8fcd696335c2\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9" Jan 21 18:06:22 crc kubenswrapper[4808]: I0121 18:06:22.334047 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9" Jan 21 18:06:22 crc kubenswrapper[4808]: I0121 18:06:22.524546 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9"] Jan 21 18:06:22 crc kubenswrapper[4808]: W0121 18:06:22.527587 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61b09e55_6f9f_4dd0_a174_8fcd696335c2.slice/crio-5195b501ab7b97ba7648d2652f604d8a757f4e1f6c039fa8bb02fd5560575493 WatchSource:0}: Error finding container 5195b501ab7b97ba7648d2652f604d8a757f4e1f6c039fa8bb02fd5560575493: Status 404 returned error can't find the container with id 5195b501ab7b97ba7648d2652f604d8a757f4e1f6c039fa8bb02fd5560575493 Jan 21 18:06:23 crc kubenswrapper[4808]: I0121 18:06:23.113202 4808 generic.go:334] "Generic (PLEG): container finished" podID="8935b136-3537-4e51-ae90-9c3729258689" containerID="d4b0faf92f9aa608eb22842b99877514ce45ccf670276a7db21fdf73944afcf4" exitCode=0 Jan 21 18:06:23 crc kubenswrapper[4808]: I0121 18:06:23.113302 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm" event={"ID":"8935b136-3537-4e51-ae90-9c3729258689","Type":"ContainerDied","Data":"d4b0faf92f9aa608eb22842b99877514ce45ccf670276a7db21fdf73944afcf4"} Jan 21 18:06:23 crc kubenswrapper[4808]: I0121 18:06:23.115320 4808 generic.go:334] "Generic (PLEG): container finished" podID="61b09e55-6f9f-4dd0-a174-8fcd696335c2" containerID="9307631d17999672e2ee29309a0fbb4b7753abcc8d74d99417b5a976cc636ba2" exitCode=0 Jan 21 18:06:23 crc kubenswrapper[4808]: I0121 18:06:23.115369 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9" event={"ID":"61b09e55-6f9f-4dd0-a174-8fcd696335c2","Type":"ContainerDied","Data":"9307631d17999672e2ee29309a0fbb4b7753abcc8d74d99417b5a976cc636ba2"} Jan 21 18:06:23 crc kubenswrapper[4808]: I0121 18:06:23.115405 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9" event={"ID":"61b09e55-6f9f-4dd0-a174-8fcd696335c2","Type":"ContainerStarted","Data":"5195b501ab7b97ba7648d2652f604d8a757f4e1f6c039fa8bb02fd5560575493"} Jan 21 18:06:26 crc kubenswrapper[4808]: I0121 18:06:26.136780 4808 generic.go:334] "Generic (PLEG): container finished" podID="8935b136-3537-4e51-ae90-9c3729258689" containerID="b26ff958f157dfcf83beda84d1b0857cafb7b2b366b50067f8fb9e30eaaaaab6" exitCode=0 Jan 21 18:06:26 crc kubenswrapper[4808]: I0121 18:06:26.136826 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm" event={"ID":"8935b136-3537-4e51-ae90-9c3729258689","Type":"ContainerDied","Data":"b26ff958f157dfcf83beda84d1b0857cafb7b2b366b50067f8fb9e30eaaaaab6"} Jan 21 18:06:27 crc kubenswrapper[4808]: I0121 18:06:27.147449 4808 generic.go:334] "Generic (PLEG): container finished" podID="8935b136-3537-4e51-ae90-9c3729258689" containerID="b3dbb6d0400c00b0d42e94b802f3b34ea139c23ff73b242d6e10cd823f02f0cf" exitCode=0 Jan 21 18:06:27 crc kubenswrapper[4808]: I0121 18:06:27.147552 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm" event={"ID":"8935b136-3537-4e51-ae90-9c3729258689","Type":"ContainerDied","Data":"b3dbb6d0400c00b0d42e94b802f3b34ea139c23ff73b242d6e10cd823f02f0cf"} Jan 21 18:06:28 crc kubenswrapper[4808]: I0121 18:06:28.396037 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm" Jan 21 18:06:28 crc kubenswrapper[4808]: I0121 18:06:28.564563 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8935b136-3537-4e51-ae90-9c3729258689-util\") pod \"8935b136-3537-4e51-ae90-9c3729258689\" (UID: \"8935b136-3537-4e51-ae90-9c3729258689\") " Jan 21 18:06:28 crc kubenswrapper[4808]: I0121 18:06:28.564694 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6fjp\" (UniqueName: \"kubernetes.io/projected/8935b136-3537-4e51-ae90-9c3729258689-kube-api-access-p6fjp\") pod \"8935b136-3537-4e51-ae90-9c3729258689\" (UID: \"8935b136-3537-4e51-ae90-9c3729258689\") " Jan 21 18:06:28 crc kubenswrapper[4808]: I0121 18:06:28.564738 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8935b136-3537-4e51-ae90-9c3729258689-bundle\") pod \"8935b136-3537-4e51-ae90-9c3729258689\" (UID: \"8935b136-3537-4e51-ae90-9c3729258689\") " Jan 21 18:06:28 crc kubenswrapper[4808]: I0121 18:06:28.565183 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8935b136-3537-4e51-ae90-9c3729258689-bundle" (OuterVolumeSpecName: "bundle") pod "8935b136-3537-4e51-ae90-9c3729258689" (UID: "8935b136-3537-4e51-ae90-9c3729258689"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:06:28 crc kubenswrapper[4808]: I0121 18:06:28.570614 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8935b136-3537-4e51-ae90-9c3729258689-kube-api-access-p6fjp" (OuterVolumeSpecName: "kube-api-access-p6fjp") pod "8935b136-3537-4e51-ae90-9c3729258689" (UID: "8935b136-3537-4e51-ae90-9c3729258689"). InnerVolumeSpecName "kube-api-access-p6fjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:06:28 crc kubenswrapper[4808]: I0121 18:06:28.575397 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8935b136-3537-4e51-ae90-9c3729258689-util" (OuterVolumeSpecName: "util") pod "8935b136-3537-4e51-ae90-9c3729258689" (UID: "8935b136-3537-4e51-ae90-9c3729258689"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:06:28 crc kubenswrapper[4808]: I0121 18:06:28.665992 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6fjp\" (UniqueName: \"kubernetes.io/projected/8935b136-3537-4e51-ae90-9c3729258689-kube-api-access-p6fjp\") on node \"crc\" DevicePath \"\"" Jan 21 18:06:28 crc kubenswrapper[4808]: I0121 18:06:28.666041 4808 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8935b136-3537-4e51-ae90-9c3729258689-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 18:06:28 crc kubenswrapper[4808]: I0121 18:06:28.666053 4808 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8935b136-3537-4e51-ae90-9c3729258689-util\") on node \"crc\" DevicePath \"\"" Jan 21 18:06:29 crc kubenswrapper[4808]: I0121 18:06:29.161441 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm" event={"ID":"8935b136-3537-4e51-ae90-9c3729258689","Type":"ContainerDied","Data":"2b50cfb238a2d21aedf9f21c91efcf783990badacc259892f32c009b3115fe8a"} Jan 21 18:06:29 crc kubenswrapper[4808]: I0121 18:06:29.161668 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b50cfb238a2d21aedf9f21c91efcf783990badacc259892f32c009b3115fe8a" Jan 21 18:06:29 crc kubenswrapper[4808]: I0121 18:06:29.161512 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm" Jan 21 18:06:29 crc kubenswrapper[4808]: I0121 18:06:29.600438 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:06:29 crc kubenswrapper[4808]: I0121 18:06:29.600517 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:06:29 crc kubenswrapper[4808]: I0121 18:06:29.600573 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 18:06:29 crc kubenswrapper[4808]: I0121 18:06:29.601229 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"00f0bd827addaebc04ffc5544dc603b3fe979bfa986819c93eca5d866ec75282"} pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 18:06:29 crc kubenswrapper[4808]: I0121 18:06:29.601312 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" containerID="cri-o://00f0bd827addaebc04ffc5544dc603b3fe979bfa986819c93eca5d866ec75282" gracePeriod=600 Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.177896 4808 generic.go:334] "Generic (PLEG): container finished" podID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerID="00f0bd827addaebc04ffc5544dc603b3fe979bfa986819c93eca5d866ec75282" exitCode=0 Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.177975 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerDied","Data":"00f0bd827addaebc04ffc5544dc603b3fe979bfa986819c93eca5d866ec75282"} Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.178403 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerStarted","Data":"5fd5f938b5c81db84b9cd3776775f788a436e3c24fc43c86adbeb98927f5200d"} Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.178429 4808 scope.go:117] "RemoveContainer" containerID="b1f94a9dbae93a06cda5ec2fa475694e62e99f13f0c15187ffea4d991182c486" Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.181819 4808 generic.go:334] "Generic (PLEG): container finished" podID="61b09e55-6f9f-4dd0-a174-8fcd696335c2" containerID="e16f9eee880a49b6fd8b2c5b923f061197100a4b9563f5638e4343aa6553a330" exitCode=0 Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.181874 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9" event={"ID":"61b09e55-6f9f-4dd0-a174-8fcd696335c2","Type":"ContainerDied","Data":"e16f9eee880a49b6fd8b2c5b923f061197100a4b9563f5638e4343aa6553a330"} Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.220589 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d"] Jan 21 18:06:30 crc kubenswrapper[4808]: E0121 18:06:30.220874 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8935b136-3537-4e51-ae90-9c3729258689" containerName="pull" Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.220897 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8935b136-3537-4e51-ae90-9c3729258689" containerName="pull" Jan 21 18:06:30 crc kubenswrapper[4808]: E0121 18:06:30.220909 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8935b136-3537-4e51-ae90-9c3729258689" containerName="extract" Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.220916 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8935b136-3537-4e51-ae90-9c3729258689" containerName="extract" Jan 21 18:06:30 crc kubenswrapper[4808]: E0121 18:06:30.220933 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8935b136-3537-4e51-ae90-9c3729258689" containerName="util" Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.220941 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8935b136-3537-4e51-ae90-9c3729258689" containerName="util" Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.221289 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="8935b136-3537-4e51-ae90-9c3729258689" containerName="extract" Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.222233 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d" Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.243367 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d"] Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.391605 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a8baaf63-dc73-4834-870d-bb3c7cd44ede-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d\" (UID: \"a8baaf63-dc73-4834-870d-bb3c7cd44ede\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d" Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.392100 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a8baaf63-dc73-4834-870d-bb3c7cd44ede-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d\" (UID: \"a8baaf63-dc73-4834-870d-bb3c7cd44ede\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d" Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.392148 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4mcd\" (UniqueName: \"kubernetes.io/projected/a8baaf63-dc73-4834-870d-bb3c7cd44ede-kube-api-access-h4mcd\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d\" (UID: \"a8baaf63-dc73-4834-870d-bb3c7cd44ede\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d" Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.493350 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a8baaf63-dc73-4834-870d-bb3c7cd44ede-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d\" (UID: \"a8baaf63-dc73-4834-870d-bb3c7cd44ede\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d" Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.493578 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a8baaf63-dc73-4834-870d-bb3c7cd44ede-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d\" (UID: \"a8baaf63-dc73-4834-870d-bb3c7cd44ede\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d" Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.493665 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4mcd\" (UniqueName: \"kubernetes.io/projected/a8baaf63-dc73-4834-870d-bb3c7cd44ede-kube-api-access-h4mcd\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d\" (UID: \"a8baaf63-dc73-4834-870d-bb3c7cd44ede\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d" Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.494082 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a8baaf63-dc73-4834-870d-bb3c7cd44ede-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d\" (UID: \"a8baaf63-dc73-4834-870d-bb3c7cd44ede\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d" Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.494467 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a8baaf63-dc73-4834-870d-bb3c7cd44ede-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d\" (UID: \"a8baaf63-dc73-4834-870d-bb3c7cd44ede\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d" Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.524277 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4mcd\" (UniqueName: \"kubernetes.io/projected/a8baaf63-dc73-4834-870d-bb3c7cd44ede-kube-api-access-h4mcd\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d\" (UID: \"a8baaf63-dc73-4834-870d-bb3c7cd44ede\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d" Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.574304 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d" Jan 21 18:06:30 crc kubenswrapper[4808]: I0121 18:06:30.871273 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d"] Jan 21 18:06:30 crc kubenswrapper[4808]: W0121 18:06:30.882548 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8baaf63_dc73_4834_870d_bb3c7cd44ede.slice/crio-9120f8cb4e49da69be3bce4d5e6133271cb8d933e70b74f6406c104ceb20bff4 WatchSource:0}: Error finding container 9120f8cb4e49da69be3bce4d5e6133271cb8d933e70b74f6406c104ceb20bff4: Status 404 returned error can't find the container with id 9120f8cb4e49da69be3bce4d5e6133271cb8d933e70b74f6406c104ceb20bff4 Jan 21 18:06:31 crc kubenswrapper[4808]: I0121 18:06:31.192148 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d" event={"ID":"a8baaf63-dc73-4834-870d-bb3c7cd44ede","Type":"ContainerStarted","Data":"9120f8cb4e49da69be3bce4d5e6133271cb8d933e70b74f6406c104ceb20bff4"} Jan 21 18:06:31 crc kubenswrapper[4808]: I0121 18:06:31.194857 4808 generic.go:334] "Generic (PLEG): container finished" podID="61b09e55-6f9f-4dd0-a174-8fcd696335c2" containerID="4d67c8612c111e0cec0642af38eff0a3ea186ff0a1225c0f75c4b0678ea8554b" exitCode=0 Jan 21 18:06:31 crc kubenswrapper[4808]: I0121 18:06:31.194897 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9" event={"ID":"61b09e55-6f9f-4dd0-a174-8fcd696335c2","Type":"ContainerDied","Data":"4d67c8612c111e0cec0642af38eff0a3ea186ff0a1225c0f75c4b0678ea8554b"} Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.162319 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-6mhk8"] Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.163354 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6mhk8" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.165612 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.170882 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.170913 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-4qg9m" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.181474 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-6mhk8"] Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.202077 4808 generic.go:334] "Generic (PLEG): container finished" podID="a8baaf63-dc73-4834-870d-bb3c7cd44ede" containerID="3a33ca84edbb89e21214ec17ee5f33e14d3bc69c0cdf470211306c293645a8ea" exitCode=0 Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.202172 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d" event={"ID":"a8baaf63-dc73-4834-870d-bb3c7cd44ede","Type":"ContainerDied","Data":"3a33ca84edbb89e21214ec17ee5f33e14d3bc69c0cdf470211306c293645a8ea"} Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.215530 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tvhn\" (UniqueName: \"kubernetes.io/projected/14e13b58-9524-4ad9-94db-848ef5698394-kube-api-access-2tvhn\") pod \"obo-prometheus-operator-68bc856cb9-6mhk8\" (UID: \"14e13b58-9524-4ad9-94db-848ef5698394\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6mhk8" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.292171 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr"] Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.293024 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.296048 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-9fqhn" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.296505 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.307753 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr"] Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.316668 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tvhn\" (UniqueName: \"kubernetes.io/projected/14e13b58-9524-4ad9-94db-848ef5698394-kube-api-access-2tvhn\") pod \"obo-prometheus-operator-68bc856cb9-6mhk8\" (UID: \"14e13b58-9524-4ad9-94db-848ef5698394\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6mhk8" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.332517 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-dss74"] Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.333260 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-dss74" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.334338 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-dss74"] Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.346597 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tvhn\" (UniqueName: \"kubernetes.io/projected/14e13b58-9524-4ad9-94db-848ef5698394-kube-api-access-2tvhn\") pod \"obo-prometheus-operator-68bc856cb9-6mhk8\" (UID: \"14e13b58-9524-4ad9-94db-848ef5698394\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6mhk8" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.418581 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a0ebf276-ed9c-4cb7-9423-0089c9e08533-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-59957c94cc-dss74\" (UID: \"a0ebf276-ed9c-4cb7-9423-0089c9e08533\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-dss74" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.419013 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5e732623-a8de-4909-850b-6728c2db0689-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr\" (UID: \"5e732623-a8de-4909-850b-6728c2db0689\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.419059 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5e732623-a8de-4909-850b-6728c2db0689-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr\" (UID: \"5e732623-a8de-4909-850b-6728c2db0689\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.419122 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a0ebf276-ed9c-4cb7-9423-0089c9e08533-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-59957c94cc-dss74\" (UID: \"a0ebf276-ed9c-4cb7-9423-0089c9e08533\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-dss74" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.446102 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-fh9px"] Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.446879 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-fh9px" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.449367 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.449367 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-lb5gn" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.461941 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-fh9px"] Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.480971 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6mhk8" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.523379 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a0ebf276-ed9c-4cb7-9423-0089c9e08533-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-59957c94cc-dss74\" (UID: \"a0ebf276-ed9c-4cb7-9423-0089c9e08533\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-dss74" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.523466 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a0ebf276-ed9c-4cb7-9423-0089c9e08533-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-59957c94cc-dss74\" (UID: \"a0ebf276-ed9c-4cb7-9423-0089c9e08533\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-dss74" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.523501 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5e732623-a8de-4909-850b-6728c2db0689-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr\" (UID: \"5e732623-a8de-4909-850b-6728c2db0689\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.523554 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5e732623-a8de-4909-850b-6728c2db0689-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr\" (UID: \"5e732623-a8de-4909-850b-6728c2db0689\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.529820 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a0ebf276-ed9c-4cb7-9423-0089c9e08533-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-59957c94cc-dss74\" (UID: \"a0ebf276-ed9c-4cb7-9423-0089c9e08533\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-dss74" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.531118 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5e732623-a8de-4909-850b-6728c2db0689-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr\" (UID: \"5e732623-a8de-4909-850b-6728c2db0689\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.531118 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a0ebf276-ed9c-4cb7-9423-0089c9e08533-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-59957c94cc-dss74\" (UID: \"a0ebf276-ed9c-4cb7-9423-0089c9e08533\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-dss74" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.535779 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5e732623-a8de-4909-850b-6728c2db0689-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr\" (UID: \"5e732623-a8de-4909-850b-6728c2db0689\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.569065 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.598024 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-lkgq6"] Jan 21 18:06:32 crc kubenswrapper[4808]: E0121 18:06:32.599336 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61b09e55-6f9f-4dd0-a174-8fcd696335c2" containerName="util" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.599433 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="61b09e55-6f9f-4dd0-a174-8fcd696335c2" containerName="util" Jan 21 18:06:32 crc kubenswrapper[4808]: E0121 18:06:32.599496 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61b09e55-6f9f-4dd0-a174-8fcd696335c2" containerName="pull" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.599544 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="61b09e55-6f9f-4dd0-a174-8fcd696335c2" containerName="pull" Jan 21 18:06:32 crc kubenswrapper[4808]: E0121 18:06:32.599594 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61b09e55-6f9f-4dd0-a174-8fcd696335c2" containerName="extract" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.599658 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="61b09e55-6f9f-4dd0-a174-8fcd696335c2" containerName="extract" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.599827 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="61b09e55-6f9f-4dd0-a174-8fcd696335c2" containerName="extract" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.600339 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-lkgq6" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.602711 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-hxhgl" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.612660 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.626226 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hspkp\" (UniqueName: \"kubernetes.io/projected/a54735e0-9845-41e8-a765-cb6a4d04943c-kube-api-access-hspkp\") pod \"observability-operator-59bdc8b94-fh9px\" (UID: \"a54735e0-9845-41e8-a765-cb6a4d04943c\") " pod="openshift-operators/observability-operator-59bdc8b94-fh9px" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.626348 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/a54735e0-9845-41e8-a765-cb6a4d04943c-observability-operator-tls\") pod \"observability-operator-59bdc8b94-fh9px\" (UID: \"a54735e0-9845-41e8-a765-cb6a4d04943c\") " pod="openshift-operators/observability-operator-59bdc8b94-fh9px" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.626368 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/f24a8af6-f286-470e-a41e-86d9cdd38963-openshift-service-ca\") pod \"perses-operator-5bf474d74f-lkgq6\" (UID: \"f24a8af6-f286-470e-a41e-86d9cdd38963\") " pod="openshift-operators/perses-operator-5bf474d74f-lkgq6" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.626491 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74mxh\" (UniqueName: \"kubernetes.io/projected/f24a8af6-f286-470e-a41e-86d9cdd38963-kube-api-access-74mxh\") pod \"perses-operator-5bf474d74f-lkgq6\" (UID: \"f24a8af6-f286-470e-a41e-86d9cdd38963\") " pod="openshift-operators/perses-operator-5bf474d74f-lkgq6" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.627853 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-lkgq6"] Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.679907 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-dss74" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.727142 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnshv\" (UniqueName: \"kubernetes.io/projected/61b09e55-6f9f-4dd0-a174-8fcd696335c2-kube-api-access-tnshv\") pod \"61b09e55-6f9f-4dd0-a174-8fcd696335c2\" (UID: \"61b09e55-6f9f-4dd0-a174-8fcd696335c2\") " Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.727571 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/61b09e55-6f9f-4dd0-a174-8fcd696335c2-bundle\") pod \"61b09e55-6f9f-4dd0-a174-8fcd696335c2\" (UID: \"61b09e55-6f9f-4dd0-a174-8fcd696335c2\") " Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.727599 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/61b09e55-6f9f-4dd0-a174-8fcd696335c2-util\") pod \"61b09e55-6f9f-4dd0-a174-8fcd696335c2\" (UID: \"61b09e55-6f9f-4dd0-a174-8fcd696335c2\") " Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.729689 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hspkp\" (UniqueName: \"kubernetes.io/projected/a54735e0-9845-41e8-a765-cb6a4d04943c-kube-api-access-hspkp\") pod \"observability-operator-59bdc8b94-fh9px\" (UID: \"a54735e0-9845-41e8-a765-cb6a4d04943c\") " pod="openshift-operators/observability-operator-59bdc8b94-fh9px" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.729780 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/a54735e0-9845-41e8-a765-cb6a4d04943c-observability-operator-tls\") pod \"observability-operator-59bdc8b94-fh9px\" (UID: \"a54735e0-9845-41e8-a765-cb6a4d04943c\") " pod="openshift-operators/observability-operator-59bdc8b94-fh9px" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.729814 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/f24a8af6-f286-470e-a41e-86d9cdd38963-openshift-service-ca\") pod \"perses-operator-5bf474d74f-lkgq6\" (UID: \"f24a8af6-f286-470e-a41e-86d9cdd38963\") " pod="openshift-operators/perses-operator-5bf474d74f-lkgq6" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.729891 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74mxh\" (UniqueName: \"kubernetes.io/projected/f24a8af6-f286-470e-a41e-86d9cdd38963-kube-api-access-74mxh\") pod \"perses-operator-5bf474d74f-lkgq6\" (UID: \"f24a8af6-f286-470e-a41e-86d9cdd38963\") " pod="openshift-operators/perses-operator-5bf474d74f-lkgq6" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.730035 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61b09e55-6f9f-4dd0-a174-8fcd696335c2-bundle" (OuterVolumeSpecName: "bundle") pod "61b09e55-6f9f-4dd0-a174-8fcd696335c2" (UID: "61b09e55-6f9f-4dd0-a174-8fcd696335c2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.740726 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61b09e55-6f9f-4dd0-a174-8fcd696335c2-util" (OuterVolumeSpecName: "util") pod "61b09e55-6f9f-4dd0-a174-8fcd696335c2" (UID: "61b09e55-6f9f-4dd0-a174-8fcd696335c2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.745664 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61b09e55-6f9f-4dd0-a174-8fcd696335c2-kube-api-access-tnshv" (OuterVolumeSpecName: "kube-api-access-tnshv") pod "61b09e55-6f9f-4dd0-a174-8fcd696335c2" (UID: "61b09e55-6f9f-4dd0-a174-8fcd696335c2"). InnerVolumeSpecName "kube-api-access-tnshv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.747105 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/f24a8af6-f286-470e-a41e-86d9cdd38963-openshift-service-ca\") pod \"perses-operator-5bf474d74f-lkgq6\" (UID: \"f24a8af6-f286-470e-a41e-86d9cdd38963\") " pod="openshift-operators/perses-operator-5bf474d74f-lkgq6" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.750106 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hspkp\" (UniqueName: \"kubernetes.io/projected/a54735e0-9845-41e8-a765-cb6a4d04943c-kube-api-access-hspkp\") pod \"observability-operator-59bdc8b94-fh9px\" (UID: \"a54735e0-9845-41e8-a765-cb6a4d04943c\") " pod="openshift-operators/observability-operator-59bdc8b94-fh9px" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.756268 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/a54735e0-9845-41e8-a765-cb6a4d04943c-observability-operator-tls\") pod \"observability-operator-59bdc8b94-fh9px\" (UID: \"a54735e0-9845-41e8-a765-cb6a4d04943c\") " pod="openshift-operators/observability-operator-59bdc8b94-fh9px" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.772422 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-fh9px" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.797883 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-6mhk8"] Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.804103 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74mxh\" (UniqueName: \"kubernetes.io/projected/f24a8af6-f286-470e-a41e-86d9cdd38963-kube-api-access-74mxh\") pod \"perses-operator-5bf474d74f-lkgq6\" (UID: \"f24a8af6-f286-470e-a41e-86d9cdd38963\") " pod="openshift-operators/perses-operator-5bf474d74f-lkgq6" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.830704 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnshv\" (UniqueName: \"kubernetes.io/projected/61b09e55-6f9f-4dd0-a174-8fcd696335c2-kube-api-access-tnshv\") on node \"crc\" DevicePath \"\"" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.830722 4808 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/61b09e55-6f9f-4dd0-a174-8fcd696335c2-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.830731 4808 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/61b09e55-6f9f-4dd0-a174-8fcd696335c2-util\") on node \"crc\" DevicePath \"\"" Jan 21 18:06:32 crc kubenswrapper[4808]: I0121 18:06:32.938566 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-lkgq6" Jan 21 18:06:33 crc kubenswrapper[4808]: I0121 18:06:33.040188 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr"] Jan 21 18:06:33 crc kubenswrapper[4808]: I0121 18:06:33.116776 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-dss74"] Jan 21 18:06:33 crc kubenswrapper[4808]: W0121 18:06:33.125404 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0ebf276_ed9c_4cb7_9423_0089c9e08533.slice/crio-11c71a0e142570db56e6f7f064743d3c676a620851364bfc6b6c8e474267f7dd WatchSource:0}: Error finding container 11c71a0e142570db56e6f7f064743d3c676a620851364bfc6b6c8e474267f7dd: Status 404 returned error can't find the container with id 11c71a0e142570db56e6f7f064743d3c676a620851364bfc6b6c8e474267f7dd Jan 21 18:06:33 crc kubenswrapper[4808]: I0121 18:06:33.184303 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-fh9px"] Jan 21 18:06:33 crc kubenswrapper[4808]: I0121 18:06:33.225393 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-dss74" event={"ID":"a0ebf276-ed9c-4cb7-9423-0089c9e08533","Type":"ContainerStarted","Data":"11c71a0e142570db56e6f7f064743d3c676a620851364bfc6b6c8e474267f7dd"} Jan 21 18:06:33 crc kubenswrapper[4808]: I0121 18:06:33.230553 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6mhk8" event={"ID":"14e13b58-9524-4ad9-94db-848ef5698394","Type":"ContainerStarted","Data":"aa7f03e32e643a2543a4c255dd8a14b6755d703f5c0cf532fdb1da7218dc8943"} Jan 21 18:06:33 crc kubenswrapper[4808]: I0121 18:06:33.232744 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-fh9px" event={"ID":"a54735e0-9845-41e8-a765-cb6a4d04943c","Type":"ContainerStarted","Data":"2c2e83583c90fb4670f42dcd4e172e9596ab7724159a896f3a3599ed0ef8cfc1"} Jan 21 18:06:33 crc kubenswrapper[4808]: I0121 18:06:33.235879 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9" event={"ID":"61b09e55-6f9f-4dd0-a174-8fcd696335c2","Type":"ContainerDied","Data":"5195b501ab7b97ba7648d2652f604d8a757f4e1f6c039fa8bb02fd5560575493"} Jan 21 18:06:33 crc kubenswrapper[4808]: I0121 18:06:33.235941 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5195b501ab7b97ba7648d2652f604d8a757f4e1f6c039fa8bb02fd5560575493" Jan 21 18:06:33 crc kubenswrapper[4808]: I0121 18:06:33.236048 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9" Jan 21 18:06:33 crc kubenswrapper[4808]: I0121 18:06:33.237837 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr" event={"ID":"5e732623-a8de-4909-850b-6728c2db0689","Type":"ContainerStarted","Data":"365c0776f1deaf6fe6d95bc75509ff8bb0bc9e0bb2ea79eba74b73acf8357721"} Jan 21 18:06:33 crc kubenswrapper[4808]: I0121 18:06:33.241629 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-lkgq6"] Jan 21 18:06:33 crc kubenswrapper[4808]: W0121 18:06:33.258237 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf24a8af6_f286_470e_a41e_86d9cdd38963.slice/crio-37458a94fb82e38105b3f0d3e3d10e633a5693ad0def22978d2a3ac3db77ab94 WatchSource:0}: Error finding container 37458a94fb82e38105b3f0d3e3d10e633a5693ad0def22978d2a3ac3db77ab94: Status 404 returned error can't find the container with id 37458a94fb82e38105b3f0d3e3d10e633a5693ad0def22978d2a3ac3db77ab94 Jan 21 18:06:34 crc kubenswrapper[4808]: I0121 18:06:34.247480 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-lkgq6" event={"ID":"f24a8af6-f286-470e-a41e-86d9cdd38963","Type":"ContainerStarted","Data":"37458a94fb82e38105b3f0d3e3d10e633a5693ad0def22978d2a3ac3db77ab94"} Jan 21 18:06:40 crc kubenswrapper[4808]: I0121 18:06:40.954842 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-g47xc"] Jan 21 18:06:40 crc kubenswrapper[4808]: I0121 18:06:40.956393 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-g47xc" Jan 21 18:06:40 crc kubenswrapper[4808]: I0121 18:06:40.962842 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"openshift-service-ca.crt" Jan 21 18:06:40 crc kubenswrapper[4808]: I0121 18:06:40.963220 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"kube-root-ca.crt" Jan 21 18:06:40 crc kubenswrapper[4808]: I0121 18:06:40.965332 4808 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"interconnect-operator-dockercfg-n6jk5" Jan 21 18:06:41 crc kubenswrapper[4808]: I0121 18:06:41.042027 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-g47xc"] Jan 21 18:06:41 crc kubenswrapper[4808]: I0121 18:06:41.060997 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pj5qs\" (UniqueName: \"kubernetes.io/projected/47dc84b9-86ae-4fbb-af34-c628708ac495-kube-api-access-pj5qs\") pod \"interconnect-operator-5bb49f789d-g47xc\" (UID: \"47dc84b9-86ae-4fbb-af34-c628708ac495\") " pod="service-telemetry/interconnect-operator-5bb49f789d-g47xc" Jan 21 18:06:41 crc kubenswrapper[4808]: I0121 18:06:41.162233 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pj5qs\" (UniqueName: \"kubernetes.io/projected/47dc84b9-86ae-4fbb-af34-c628708ac495-kube-api-access-pj5qs\") pod \"interconnect-operator-5bb49f789d-g47xc\" (UID: \"47dc84b9-86ae-4fbb-af34-c628708ac495\") " pod="service-telemetry/interconnect-operator-5bb49f789d-g47xc" Jan 21 18:06:41 crc kubenswrapper[4808]: I0121 18:06:41.192329 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pj5qs\" (UniqueName: \"kubernetes.io/projected/47dc84b9-86ae-4fbb-af34-c628708ac495-kube-api-access-pj5qs\") pod \"interconnect-operator-5bb49f789d-g47xc\" (UID: \"47dc84b9-86ae-4fbb-af34-c628708ac495\") " pod="service-telemetry/interconnect-operator-5bb49f789d-g47xc" Jan 21 18:06:41 crc kubenswrapper[4808]: I0121 18:06:41.328167 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-g47xc" Jan 21 18:06:43 crc kubenswrapper[4808]: I0121 18:06:43.491117 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elastic-operator-7bddffb8f9-vlbqx"] Jan 21 18:06:43 crc kubenswrapper[4808]: I0121 18:06:43.492154 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-7bddffb8f9-vlbqx" Jan 21 18:06:43 crc kubenswrapper[4808]: I0121 18:06:43.494372 4808 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-service-cert" Jan 21 18:06:43 crc kubenswrapper[4808]: I0121 18:06:43.499414 4808 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-dockercfg-m7gp5" Jan 21 18:06:43 crc kubenswrapper[4808]: I0121 18:06:43.503725 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x696f\" (UniqueName: \"kubernetes.io/projected/2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc-kube-api-access-x696f\") pod \"elastic-operator-7bddffb8f9-vlbqx\" (UID: \"2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc\") " pod="service-telemetry/elastic-operator-7bddffb8f9-vlbqx" Jan 21 18:06:43 crc kubenswrapper[4808]: I0121 18:06:43.503799 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc-webhook-cert\") pod \"elastic-operator-7bddffb8f9-vlbqx\" (UID: \"2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc\") " pod="service-telemetry/elastic-operator-7bddffb8f9-vlbqx" Jan 21 18:06:43 crc kubenswrapper[4808]: I0121 18:06:43.503835 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc-apiservice-cert\") pod \"elastic-operator-7bddffb8f9-vlbqx\" (UID: \"2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc\") " pod="service-telemetry/elastic-operator-7bddffb8f9-vlbqx" Jan 21 18:06:43 crc kubenswrapper[4808]: I0121 18:06:43.508570 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-7bddffb8f9-vlbqx"] Jan 21 18:06:43 crc kubenswrapper[4808]: I0121 18:06:43.605461 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x696f\" (UniqueName: \"kubernetes.io/projected/2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc-kube-api-access-x696f\") pod \"elastic-operator-7bddffb8f9-vlbqx\" (UID: \"2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc\") " pod="service-telemetry/elastic-operator-7bddffb8f9-vlbqx" Jan 21 18:06:43 crc kubenswrapper[4808]: I0121 18:06:43.605547 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc-webhook-cert\") pod \"elastic-operator-7bddffb8f9-vlbqx\" (UID: \"2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc\") " pod="service-telemetry/elastic-operator-7bddffb8f9-vlbqx" Jan 21 18:06:43 crc kubenswrapper[4808]: I0121 18:06:43.605575 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc-apiservice-cert\") pod \"elastic-operator-7bddffb8f9-vlbqx\" (UID: \"2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc\") " pod="service-telemetry/elastic-operator-7bddffb8f9-vlbqx" Jan 21 18:06:43 crc kubenswrapper[4808]: I0121 18:06:43.612287 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc-webhook-cert\") pod \"elastic-operator-7bddffb8f9-vlbqx\" (UID: \"2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc\") " pod="service-telemetry/elastic-operator-7bddffb8f9-vlbqx" Jan 21 18:06:43 crc kubenswrapper[4808]: I0121 18:06:43.612331 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc-apiservice-cert\") pod \"elastic-operator-7bddffb8f9-vlbqx\" (UID: \"2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc\") " pod="service-telemetry/elastic-operator-7bddffb8f9-vlbqx" Jan 21 18:06:43 crc kubenswrapper[4808]: I0121 18:06:43.631113 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x696f\" (UniqueName: \"kubernetes.io/projected/2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc-kube-api-access-x696f\") pod \"elastic-operator-7bddffb8f9-vlbqx\" (UID: \"2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc\") " pod="service-telemetry/elastic-operator-7bddffb8f9-vlbqx" Jan 21 18:06:43 crc kubenswrapper[4808]: I0121 18:06:43.812113 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-7bddffb8f9-vlbqx" Jan 21 18:06:50 crc kubenswrapper[4808]: E0121 18:06:50.492090 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:2ecf763b02048d2cf4c17967a7b2cacc7afd6af0e963a39579d876f8f4170e3c" Jan 21 18:06:50 crc kubenswrapper[4808]: E0121 18:06:50.495193 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:2ecf763b02048d2cf4c17967a7b2cacc7afd6af0e963a39579d876f8f4170e3c,Command:[],Args:[--namespace=$(NAMESPACE) --images=perses=$(RELATED_IMAGE_PERSES) --images=alertmanager=$(RELATED_IMAGE_ALERTMANAGER) --images=prometheus=$(RELATED_IMAGE_PROMETHEUS) --images=thanos=$(RELATED_IMAGE_THANOS) --images=ui-dashboards=$(RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN) --images=ui-distributed-tracing=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN) --images=ui-distributed-tracing-pf5=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5) --images=ui-distributed-tracing-pf4=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4) --images=ui-logging=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN) --images=ui-logging-pf4=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4) --images=ui-troubleshooting-panel=$(RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN) --images=ui-monitoring=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN) --images=ui-monitoring-pf5=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5) --images=korrel8r=$(RELATED_IMAGE_KORREL8R) --images=health-analyzer=$(RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER) --openshift.enabled=true],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:RELATED_IMAGE_ALERTMANAGER,Value:registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:dc62889b883f597de91b5389cc52c84c607247d49a807693be2f688e4703dfc3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS,Value:registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:1b555e21bba7c609111ace4380382a696d9aceeb6e9816bf9023b8f689b6c741,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_THANOS,Value:registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:a223bab813b82d698992490bbb60927f6288a83ba52d539836c250e1471f6d34,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PERSES,Value:registry.redhat.io/cluster-observability-operator/perses-rhel9@sha256:e797cdb47beef40b04da7b6d645bca3dc32e6247003c45b56b38efd9e13bf01c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-rhel9@sha256:093d2731ac848ed5fd57356b155a19d3bf7b8db96d95b09c5d0095e143f7254f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-rhel9@sha256:7d662a120305e2528acc7e9142b770b5b6a7f4932ddfcadfa4ac953935124895,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-pf5-rhel9@sha256:75465aabb0aa427a5c531a8fcde463f6d119afbcc618ebcbf6b7ee9bc8aad160,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-pf4-rhel9@sha256:dc18c8d6a4a9a0a574a57cc5082c8a9b26023bd6d69b9732892d584c1dfe5070,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-rhel9@sha256:369729978cecdc13c99ef3d179f8eb8a450a4a0cb70b63c27a55a15d1710ba27,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-pf4-rhel9@sha256:d8c7a61d147f62b204d5c5f16864386025393453c9a81ea327bbd25d7765d611,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-rhel9@sha256:b4a6eb1cc118a4334b424614959d8b7f361ddd779b3a72690ca49b0a3f26d9b8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-rhel9@sha256:21d4fff670893ba4b7fbc528cd49f8b71c8281cede9ef84f0697065bb6a7fc50,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-pf5-rhel9@sha256:12d9dbe297a1c3b9df671f21156992082bc483887d851fafe76e5d17321ff474,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KORREL8R,Value:registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:e65c37f04f6d76a0cbfe05edb3cddf6a8f14f859ee35cf3aebea8fcb991d2c19,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER,Value:registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:48e4e178c6eeaa9d5dd77a591c185a311b4b4a5caadb7199d48463123e31dc9e,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.1,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{400 -3} {} 400m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:observability-operator-tls,ReadOnly:true,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hspkp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod observability-operator-59bdc8b94-fh9px_openshift-operators(a54735e0-9845-41e8-a765-cb6a4d04943c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 21 18:06:50 crc kubenswrapper[4808]: E0121 18:06:50.499127 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/observability-operator-59bdc8b94-fh9px" podUID="a54735e0-9845-41e8-a765-cb6a4d04943c" Jan 21 18:06:50 crc kubenswrapper[4808]: I0121 18:06:50.846597 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-g47xc"] Jan 21 18:06:51 crc kubenswrapper[4808]: I0121 18:06:51.033705 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-7bddffb8f9-vlbqx"] Jan 21 18:06:51 crc kubenswrapper[4808]: W0121 18:06:51.060641 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ce84ba2_aeb7_40c7_bcb0_09aeb792b5bc.slice/crio-4bbb2033c1fb9e6fb00b38030d433886b3a9786d5ae97525da30124517e6d63e WatchSource:0}: Error finding container 4bbb2033c1fb9e6fb00b38030d433886b3a9786d5ae97525da30124517e6d63e: Status 404 returned error can't find the container with id 4bbb2033c1fb9e6fb00b38030d433886b3a9786d5ae97525da30124517e6d63e Jan 21 18:06:51 crc kubenswrapper[4808]: I0121 18:06:51.369723 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-7bddffb8f9-vlbqx" event={"ID":"2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc","Type":"ContainerStarted","Data":"4bbb2033c1fb9e6fb00b38030d433886b3a9786d5ae97525da30124517e6d63e"} Jan 21 18:06:51 crc kubenswrapper[4808]: I0121 18:06:51.372810 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6mhk8" event={"ID":"14e13b58-9524-4ad9-94db-848ef5698394","Type":"ContainerStarted","Data":"a50f4634ee23135e99465951bb78897e843cfb22230de11803f63961b4460ceb"} Jan 21 18:06:51 crc kubenswrapper[4808]: I0121 18:06:51.374662 4808 generic.go:334] "Generic (PLEG): container finished" podID="a8baaf63-dc73-4834-870d-bb3c7cd44ede" containerID="a5b724abf9fdc3ca1aefcb073d2e709a8146262fe092f611c6028d447acad9f7" exitCode=0 Jan 21 18:06:51 crc kubenswrapper[4808]: I0121 18:06:51.374729 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d" event={"ID":"a8baaf63-dc73-4834-870d-bb3c7cd44ede","Type":"ContainerDied","Data":"a5b724abf9fdc3ca1aefcb073d2e709a8146262fe092f611c6028d447acad9f7"} Jan 21 18:06:51 crc kubenswrapper[4808]: I0121 18:06:51.376672 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-lkgq6" event={"ID":"f24a8af6-f286-470e-a41e-86d9cdd38963","Type":"ContainerStarted","Data":"0d32e30d91b2abfac33324a564d8cab12ee4ba03efd36ac43d9ced824d6c728d"} Jan 21 18:06:51 crc kubenswrapper[4808]: I0121 18:06:51.376800 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-lkgq6" Jan 21 18:06:51 crc kubenswrapper[4808]: I0121 18:06:51.380738 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr" event={"ID":"5e732623-a8de-4909-850b-6728c2db0689","Type":"ContainerStarted","Data":"4115c09f82e10b1cbafeb5c6d5a92085cf5fe6fa7413d6131ae6a0a04499f128"} Jan 21 18:06:51 crc kubenswrapper[4808]: I0121 18:06:51.382283 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-g47xc" event={"ID":"47dc84b9-86ae-4fbb-af34-c628708ac495","Type":"ContainerStarted","Data":"4ab3cbb3a81a34d1c0b67c5217e0c064b5cd2396d3cf49f3f42f6d56de2c871e"} Jan 21 18:06:51 crc kubenswrapper[4808]: I0121 18:06:51.384914 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-dss74" event={"ID":"a0ebf276-ed9c-4cb7-9423-0089c9e08533","Type":"ContainerStarted","Data":"3909e00a20a994690d50448313f1c1595cbafcb5b85f78de63db83634fb95c14"} Jan 21 18:06:51 crc kubenswrapper[4808]: E0121 18:06:51.388352 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:2ecf763b02048d2cf4c17967a7b2cacc7afd6af0e963a39579d876f8f4170e3c\\\"\"" pod="openshift-operators/observability-operator-59bdc8b94-fh9px" podUID="a54735e0-9845-41e8-a765-cb6a4d04943c" Jan 21 18:06:51 crc kubenswrapper[4808]: I0121 18:06:51.408786 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6mhk8" podStartSLOduration=1.745706784 podStartE2EDuration="19.408754301s" podCreationTimestamp="2026-01-21 18:06:32 +0000 UTC" firstStartedPulling="2026-01-21 18:06:32.850426455 +0000 UTC m=+627.580937340" lastFinishedPulling="2026-01-21 18:06:50.513473972 +0000 UTC m=+645.243984857" observedRunningTime="2026-01-21 18:06:51.401208792 +0000 UTC m=+646.131719697" watchObservedRunningTime="2026-01-21 18:06:51.408754301 +0000 UTC m=+646.139265186" Jan 21 18:06:51 crc kubenswrapper[4808]: I0121 18:06:51.449356 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-dss74" podStartSLOduration=2.058079022 podStartE2EDuration="19.449337128s" podCreationTimestamp="2026-01-21 18:06:32 +0000 UTC" firstStartedPulling="2026-01-21 18:06:33.128362483 +0000 UTC m=+627.858873368" lastFinishedPulling="2026-01-21 18:06:50.519620589 +0000 UTC m=+645.250131474" observedRunningTime="2026-01-21 18:06:51.448317775 +0000 UTC m=+646.178828660" watchObservedRunningTime="2026-01-21 18:06:51.449337128 +0000 UTC m=+646.179848013" Jan 21 18:06:51 crc kubenswrapper[4808]: I0121 18:06:51.472584 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-lkgq6" podStartSLOduration=2.260165936 podStartE2EDuration="19.472559756s" podCreationTimestamp="2026-01-21 18:06:32 +0000 UTC" firstStartedPulling="2026-01-21 18:06:33.264355482 +0000 UTC m=+627.994866357" lastFinishedPulling="2026-01-21 18:06:50.476749292 +0000 UTC m=+645.207260177" observedRunningTime="2026-01-21 18:06:51.46644841 +0000 UTC m=+646.196959305" watchObservedRunningTime="2026-01-21 18:06:51.472559756 +0000 UTC m=+646.203070641" Jan 21 18:06:51 crc kubenswrapper[4808]: I0121 18:06:51.511523 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr" podStartSLOduration=2.054087893 podStartE2EDuration="19.511479136s" podCreationTimestamp="2026-01-21 18:06:32 +0000 UTC" firstStartedPulling="2026-01-21 18:06:33.05298659 +0000 UTC m=+627.783497475" lastFinishedPulling="2026-01-21 18:06:50.510377823 +0000 UTC m=+645.240888718" observedRunningTime="2026-01-21 18:06:51.507040577 +0000 UTC m=+646.237551462" watchObservedRunningTime="2026-01-21 18:06:51.511479136 +0000 UTC m=+646.241990021" Jan 21 18:06:52 crc kubenswrapper[4808]: I0121 18:06:52.394298 4808 generic.go:334] "Generic (PLEG): container finished" podID="a8baaf63-dc73-4834-870d-bb3c7cd44ede" containerID="db50862820b6065cc4d02baa874944e7d098e5100c427f56ab13f4c65fc37520" exitCode=0 Jan 21 18:06:52 crc kubenswrapper[4808]: I0121 18:06:52.394385 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d" event={"ID":"a8baaf63-dc73-4834-870d-bb3c7cd44ede","Type":"ContainerDied","Data":"db50862820b6065cc4d02baa874944e7d098e5100c427f56ab13f4c65fc37520"} Jan 21 18:06:54 crc kubenswrapper[4808]: I0121 18:06:54.796194 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d" Jan 21 18:06:54 crc kubenswrapper[4808]: I0121 18:06:54.890699 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a8baaf63-dc73-4834-870d-bb3c7cd44ede-bundle\") pod \"a8baaf63-dc73-4834-870d-bb3c7cd44ede\" (UID: \"a8baaf63-dc73-4834-870d-bb3c7cd44ede\") " Jan 21 18:06:54 crc kubenswrapper[4808]: I0121 18:06:54.890933 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4mcd\" (UniqueName: \"kubernetes.io/projected/a8baaf63-dc73-4834-870d-bb3c7cd44ede-kube-api-access-h4mcd\") pod \"a8baaf63-dc73-4834-870d-bb3c7cd44ede\" (UID: \"a8baaf63-dc73-4834-870d-bb3c7cd44ede\") " Jan 21 18:06:54 crc kubenswrapper[4808]: I0121 18:06:54.891005 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a8baaf63-dc73-4834-870d-bb3c7cd44ede-util\") pod \"a8baaf63-dc73-4834-870d-bb3c7cd44ede\" (UID: \"a8baaf63-dc73-4834-870d-bb3c7cd44ede\") " Jan 21 18:06:54 crc kubenswrapper[4808]: I0121 18:06:54.899931 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8baaf63-dc73-4834-870d-bb3c7cd44ede-bundle" (OuterVolumeSpecName: "bundle") pod "a8baaf63-dc73-4834-870d-bb3c7cd44ede" (UID: "a8baaf63-dc73-4834-870d-bb3c7cd44ede"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:06:54 crc kubenswrapper[4808]: I0121 18:06:54.901499 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8baaf63-dc73-4834-870d-bb3c7cd44ede-kube-api-access-h4mcd" (OuterVolumeSpecName: "kube-api-access-h4mcd") pod "a8baaf63-dc73-4834-870d-bb3c7cd44ede" (UID: "a8baaf63-dc73-4834-870d-bb3c7cd44ede"). InnerVolumeSpecName "kube-api-access-h4mcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:06:54 crc kubenswrapper[4808]: I0121 18:06:54.918369 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8baaf63-dc73-4834-870d-bb3c7cd44ede-util" (OuterVolumeSpecName: "util") pod "a8baaf63-dc73-4834-870d-bb3c7cd44ede" (UID: "a8baaf63-dc73-4834-870d-bb3c7cd44ede"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:06:54 crc kubenswrapper[4808]: I0121 18:06:54.992235 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4mcd\" (UniqueName: \"kubernetes.io/projected/a8baaf63-dc73-4834-870d-bb3c7cd44ede-kube-api-access-h4mcd\") on node \"crc\" DevicePath \"\"" Jan 21 18:06:54 crc kubenswrapper[4808]: I0121 18:06:54.992286 4808 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a8baaf63-dc73-4834-870d-bb3c7cd44ede-util\") on node \"crc\" DevicePath \"\"" Jan 21 18:06:54 crc kubenswrapper[4808]: I0121 18:06:54.992298 4808 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a8baaf63-dc73-4834-870d-bb3c7cd44ede-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 18:06:55 crc kubenswrapper[4808]: I0121 18:06:55.420751 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d" event={"ID":"a8baaf63-dc73-4834-870d-bb3c7cd44ede","Type":"ContainerDied","Data":"9120f8cb4e49da69be3bce4d5e6133271cb8d933e70b74f6406c104ceb20bff4"} Jan 21 18:06:55 crc kubenswrapper[4808]: I0121 18:06:55.421140 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9120f8cb4e49da69be3bce4d5e6133271cb8d933e70b74f6406c104ceb20bff4" Jan 21 18:06:55 crc kubenswrapper[4808]: I0121 18:06:55.420803 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d" Jan 21 18:06:55 crc kubenswrapper[4808]: I0121 18:06:55.430516 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-7bddffb8f9-vlbqx" event={"ID":"2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc","Type":"ContainerStarted","Data":"b5c0a033d5dcb465c32be4e5f480303bcb5592835437bd2af0206d7cc9f64773"} Jan 21 18:06:55 crc kubenswrapper[4808]: I0121 18:06:55.476097 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elastic-operator-7bddffb8f9-vlbqx" podStartSLOduration=8.730895507 podStartE2EDuration="12.476074561s" podCreationTimestamp="2026-01-21 18:06:43 +0000 UTC" firstStartedPulling="2026-01-21 18:06:51.073361778 +0000 UTC m=+645.803872663" lastFinishedPulling="2026-01-21 18:06:54.818540832 +0000 UTC m=+649.549051717" observedRunningTime="2026-01-21 18:06:55.469457253 +0000 UTC m=+650.199968138" watchObservedRunningTime="2026-01-21 18:06:55.476074561 +0000 UTC m=+650.206585436" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.098178 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 21 18:06:56 crc kubenswrapper[4808]: E0121 18:06:56.098608 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8baaf63-dc73-4834-870d-bb3c7cd44ede" containerName="util" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.098624 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8baaf63-dc73-4834-870d-bb3c7cd44ede" containerName="util" Jan 21 18:06:56 crc kubenswrapper[4808]: E0121 18:06:56.098635 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8baaf63-dc73-4834-870d-bb3c7cd44ede" containerName="pull" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.098659 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8baaf63-dc73-4834-870d-bb3c7cd44ede" containerName="pull" Jan 21 18:06:56 crc kubenswrapper[4808]: E0121 18:06:56.098690 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8baaf63-dc73-4834-870d-bb3c7cd44ede" containerName="extract" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.098700 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8baaf63-dc73-4834-870d-bb3c7cd44ede" containerName="extract" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.098907 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8baaf63-dc73-4834-870d-bb3c7cd44ede" containerName="extract" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.100159 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.102432 4808 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-remote-ca" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.102786 4808 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-http-certs-internal" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.102956 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-scripts" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.103099 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-unicast-hosts" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.103297 4808 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-dockercfg-mhxd5" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.103524 4808 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-config" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.103637 4808 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-transport-certs" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.103838 4808 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-internal-users" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.103965 4808 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-xpack-file-realm" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.170321 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.213391 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.213457 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.213491 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/a879c359-bceb-47c5-a3db-e41219bc698d-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.213520 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/a879c359-bceb-47c5-a3db-e41219bc698d-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.213557 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/a879c359-bceb-47c5-a3db-e41219bc698d-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.213577 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.213673 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.213728 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.213752 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.213778 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/a879c359-bceb-47c5-a3db-e41219bc698d-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.213889 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.213954 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.214017 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.214055 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.214121 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.317836 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.317896 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.317924 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/a879c359-bceb-47c5-a3db-e41219bc698d-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.317952 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/a879c359-bceb-47c5-a3db-e41219bc698d-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.317984 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/a879c359-bceb-47c5-a3db-e41219bc698d-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.318007 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.318080 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.318122 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.318146 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.318170 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/a879c359-bceb-47c5-a3db-e41219bc698d-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.318194 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.318216 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.318264 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.318296 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.318341 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.319710 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.319712 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.320096 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.320411 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/a879c359-bceb-47c5-a3db-e41219bc698d-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.321216 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/a879c359-bceb-47c5-a3db-e41219bc698d-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.321458 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/a879c359-bceb-47c5-a3db-e41219bc698d-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.321780 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.322381 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.324669 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.325236 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.325367 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.325847 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.325920 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/a879c359-bceb-47c5-a3db-e41219bc698d-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.325944 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.326676 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/a879c359-bceb-47c5-a3db-e41219bc698d-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"a879c359-bceb-47c5-a3db-e41219bc698d\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.473282 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:06:56 crc kubenswrapper[4808]: I0121 18:06:56.933363 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 21 18:06:56 crc kubenswrapper[4808]: W0121 18:06:56.937104 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda879c359_bceb_47c5_a3db_e41219bc698d.slice/crio-b2c38cdb82f6c8bd0650ad545daf15133604f62a80b05417b5cf4ba49750d254 WatchSource:0}: Error finding container b2c38cdb82f6c8bd0650ad545daf15133604f62a80b05417b5cf4ba49750d254: Status 404 returned error can't find the container with id b2c38cdb82f6c8bd0650ad545daf15133604f62a80b05417b5cf4ba49750d254 Jan 21 18:06:57 crc kubenswrapper[4808]: I0121 18:06:57.449918 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"a879c359-bceb-47c5-a3db-e41219bc698d","Type":"ContainerStarted","Data":"b2c38cdb82f6c8bd0650ad545daf15133604f62a80b05417b5cf4ba49750d254"} Jan 21 18:07:02 crc kubenswrapper[4808]: I0121 18:07:02.948123 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-lkgq6" Jan 21 18:07:03 crc kubenswrapper[4808]: I0121 18:07:03.490966 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-g47xc" event={"ID":"47dc84b9-86ae-4fbb-af34-c628708ac495","Type":"ContainerStarted","Data":"c30a13e57857bf3f96d1939fdc76ec3778d47074ed27fd9392c5e5937cb8372f"} Jan 21 18:07:03 crc kubenswrapper[4808]: I0121 18:07:03.526577 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/interconnect-operator-5bb49f789d-g47xc" podStartSLOduration=11.688021888 podStartE2EDuration="23.526548053s" podCreationTimestamp="2026-01-21 18:06:40 +0000 UTC" firstStartedPulling="2026-01-21 18:06:50.893397128 +0000 UTC m=+645.623908013" lastFinishedPulling="2026-01-21 18:07:02.731923293 +0000 UTC m=+657.462434178" observedRunningTime="2026-01-21 18:07:03.504154782 +0000 UTC m=+658.234665677" watchObservedRunningTime="2026-01-21 18:07:03.526548053 +0000 UTC m=+658.257058938" Jan 21 18:07:07 crc kubenswrapper[4808]: I0121 18:07:07.527396 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-fh9px" event={"ID":"a54735e0-9845-41e8-a765-cb6a4d04943c","Type":"ContainerStarted","Data":"abf7806153c5ec3bd004d108711d00942e3a7770687f5bcdcbbfc011faea8e6f"} Jan 21 18:07:07 crc kubenswrapper[4808]: I0121 18:07:07.528489 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-fh9px" Jan 21 18:07:07 crc kubenswrapper[4808]: I0121 18:07:07.552563 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-fh9px" podStartSLOduration=1.834320766 podStartE2EDuration="35.552543145s" podCreationTimestamp="2026-01-21 18:06:32 +0000 UTC" firstStartedPulling="2026-01-21 18:06:33.188008036 +0000 UTC m=+627.918518921" lastFinishedPulling="2026-01-21 18:07:06.906230415 +0000 UTC m=+661.636741300" observedRunningTime="2026-01-21 18:07:07.547438099 +0000 UTC m=+662.277948994" watchObservedRunningTime="2026-01-21 18:07:07.552543145 +0000 UTC m=+662.283054030" Jan 21 18:07:07 crc kubenswrapper[4808]: I0121 18:07:07.647731 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-fh9px" Jan 21 18:07:14 crc kubenswrapper[4808]: I0121 18:07:14.665150 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-lr6lv"] Jan 21 18:07:14 crc kubenswrapper[4808]: I0121 18:07:14.670554 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-lr6lv" Jan 21 18:07:14 crc kubenswrapper[4808]: I0121 18:07:14.674134 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Jan 21 18:07:14 crc kubenswrapper[4808]: I0121 18:07:14.674274 4808 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-sc7t8" Jan 21 18:07:14 crc kubenswrapper[4808]: I0121 18:07:14.674326 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Jan 21 18:07:14 crc kubenswrapper[4808]: I0121 18:07:14.706105 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-lr6lv"] Jan 21 18:07:14 crc kubenswrapper[4808]: I0121 18:07:14.872265 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/03edb0f1-931a-4f49-a395-054b94c16a00-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-lr6lv\" (UID: \"03edb0f1-931a-4f49-a395-054b94c16a00\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-lr6lv" Jan 21 18:07:14 crc kubenswrapper[4808]: I0121 18:07:14.872468 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq5cv\" (UniqueName: \"kubernetes.io/projected/03edb0f1-931a-4f49-a395-054b94c16a00-kube-api-access-pq5cv\") pod \"cert-manager-operator-controller-manager-5446d6888b-lr6lv\" (UID: \"03edb0f1-931a-4f49-a395-054b94c16a00\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-lr6lv" Jan 21 18:07:14 crc kubenswrapper[4808]: I0121 18:07:14.973447 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq5cv\" (UniqueName: \"kubernetes.io/projected/03edb0f1-931a-4f49-a395-054b94c16a00-kube-api-access-pq5cv\") pod \"cert-manager-operator-controller-manager-5446d6888b-lr6lv\" (UID: \"03edb0f1-931a-4f49-a395-054b94c16a00\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-lr6lv" Jan 21 18:07:14 crc kubenswrapper[4808]: I0121 18:07:14.973895 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/03edb0f1-931a-4f49-a395-054b94c16a00-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-lr6lv\" (UID: \"03edb0f1-931a-4f49-a395-054b94c16a00\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-lr6lv" Jan 21 18:07:14 crc kubenswrapper[4808]: I0121 18:07:14.975026 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/03edb0f1-931a-4f49-a395-054b94c16a00-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-lr6lv\" (UID: \"03edb0f1-931a-4f49-a395-054b94c16a00\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-lr6lv" Jan 21 18:07:14 crc kubenswrapper[4808]: I0121 18:07:14.999691 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq5cv\" (UniqueName: \"kubernetes.io/projected/03edb0f1-931a-4f49-a395-054b94c16a00-kube-api-access-pq5cv\") pod \"cert-manager-operator-controller-manager-5446d6888b-lr6lv\" (UID: \"03edb0f1-931a-4f49-a395-054b94c16a00\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-lr6lv" Jan 21 18:07:15 crc kubenswrapper[4808]: I0121 18:07:15.285960 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-lr6lv" Jan 21 18:07:18 crc kubenswrapper[4808]: I0121 18:07:18.054092 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-lr6lv"] Jan 21 18:07:18 crc kubenswrapper[4808]: W0121 18:07:18.061757 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03edb0f1_931a_4f49_a395_054b94c16a00.slice/crio-1123dcf191f0d5b292073682e06c0f911b535762c8477b54e52dd25f6f1e1f2f WatchSource:0}: Error finding container 1123dcf191f0d5b292073682e06c0f911b535762c8477b54e52dd25f6f1e1f2f: Status 404 returned error can't find the container with id 1123dcf191f0d5b292073682e06c0f911b535762c8477b54e52dd25f6f1e1f2f Jan 21 18:07:18 crc kubenswrapper[4808]: I0121 18:07:18.611421 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"a879c359-bceb-47c5-a3db-e41219bc698d","Type":"ContainerStarted","Data":"a029e1a8e1954f5e1b07320021dc4d30a48fbca7807649d04809ab2b5d484519"} Jan 21 18:07:18 crc kubenswrapper[4808]: I0121 18:07:18.614029 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-lr6lv" event={"ID":"03edb0f1-931a-4f49-a395-054b94c16a00","Type":"ContainerStarted","Data":"1123dcf191f0d5b292073682e06c0f911b535762c8477b54e52dd25f6f1e1f2f"} Jan 21 18:07:18 crc kubenswrapper[4808]: I0121 18:07:18.767765 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 21 18:07:18 crc kubenswrapper[4808]: I0121 18:07:18.801393 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 21 18:07:20 crc kubenswrapper[4808]: I0121 18:07:20.642640 4808 generic.go:334] "Generic (PLEG): container finished" podID="a879c359-bceb-47c5-a3db-e41219bc698d" containerID="a029e1a8e1954f5e1b07320021dc4d30a48fbca7807649d04809ab2b5d484519" exitCode=0 Jan 21 18:07:20 crc kubenswrapper[4808]: I0121 18:07:20.642822 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"a879c359-bceb-47c5-a3db-e41219bc698d","Type":"ContainerDied","Data":"a029e1a8e1954f5e1b07320021dc4d30a48fbca7807649d04809ab2b5d484519"} Jan 21 18:07:23 crc kubenswrapper[4808]: I0121 18:07:23.670311 4808 generic.go:334] "Generic (PLEG): container finished" podID="a879c359-bceb-47c5-a3db-e41219bc698d" containerID="ab7f71bacbff479e6cdc498b5047b562cc5c05a0fdc7b6566ddd75cedb270d4f" exitCode=0 Jan 21 18:07:23 crc kubenswrapper[4808]: I0121 18:07:23.670640 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"a879c359-bceb-47c5-a3db-e41219bc698d","Type":"ContainerDied","Data":"ab7f71bacbff479e6cdc498b5047b562cc5c05a0fdc7b6566ddd75cedb270d4f"} Jan 21 18:07:27 crc kubenswrapper[4808]: I0121 18:07:27.697771 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"a879c359-bceb-47c5-a3db-e41219bc698d","Type":"ContainerStarted","Data":"c722667e064fa6d94d6a420f0d4df3bd11a1e6fbccb82c78f2f574e3b4fc578e"} Jan 21 18:07:27 crc kubenswrapper[4808]: I0121 18:07:27.699233 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:07:27 crc kubenswrapper[4808]: I0121 18:07:27.699479 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-lr6lv" event={"ID":"03edb0f1-931a-4f49-a395-054b94c16a00","Type":"ContainerStarted","Data":"713e37de92749bd7ef9d6b7eb4e17eff51b0c7d0acc3f2a49c632a2ca32bcf72"} Jan 21 18:07:27 crc kubenswrapper[4808]: I0121 18:07:27.737633 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elasticsearch-es-default-0" podStartSLOduration=10.607001168 podStartE2EDuration="31.737610898s" podCreationTimestamp="2026-01-21 18:06:56 +0000 UTC" firstStartedPulling="2026-01-21 18:06:56.941546379 +0000 UTC m=+651.672057264" lastFinishedPulling="2026-01-21 18:07:18.072156119 +0000 UTC m=+672.802666994" observedRunningTime="2026-01-21 18:07:27.73158471 +0000 UTC m=+682.462095615" watchObservedRunningTime="2026-01-21 18:07:27.737610898 +0000 UTC m=+682.468121783" Jan 21 18:07:27 crc kubenswrapper[4808]: I0121 18:07:27.763768 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-lr6lv" podStartSLOduration=4.632622845 podStartE2EDuration="13.763748531s" podCreationTimestamp="2026-01-21 18:07:14 +0000 UTC" firstStartedPulling="2026-01-21 18:07:18.069484313 +0000 UTC m=+672.799995208" lastFinishedPulling="2026-01-21 18:07:27.200610009 +0000 UTC m=+681.931120894" observedRunningTime="2026-01-21 18:07:27.754656347 +0000 UTC m=+682.485167242" watchObservedRunningTime="2026-01-21 18:07:27.763748531 +0000 UTC m=+682.494259406" Jan 21 18:07:30 crc kubenswrapper[4808]: I0121 18:07:30.630916 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-db8jl"] Jan 21 18:07:30 crc kubenswrapper[4808]: I0121 18:07:30.632278 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-db8jl" Jan 21 18:07:30 crc kubenswrapper[4808]: I0121 18:07:30.637894 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Jan 21 18:07:30 crc kubenswrapper[4808]: I0121 18:07:30.638678 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Jan 21 18:07:30 crc kubenswrapper[4808]: I0121 18:07:30.640503 4808 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-mxf6c" Jan 21 18:07:30 crc kubenswrapper[4808]: I0121 18:07:30.651286 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-db8jl"] Jan 21 18:07:30 crc kubenswrapper[4808]: I0121 18:07:30.710300 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq72k\" (UniqueName: \"kubernetes.io/projected/d30e2538-925e-4c64-8d47-7d0b7255f69d-kube-api-access-rq72k\") pod \"cert-manager-webhook-f4fb5df64-db8jl\" (UID: \"d30e2538-925e-4c64-8d47-7d0b7255f69d\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-db8jl" Jan 21 18:07:30 crc kubenswrapper[4808]: I0121 18:07:30.710373 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d30e2538-925e-4c64-8d47-7d0b7255f69d-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-db8jl\" (UID: \"d30e2538-925e-4c64-8d47-7d0b7255f69d\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-db8jl" Jan 21 18:07:30 crc kubenswrapper[4808]: I0121 18:07:30.811594 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq72k\" (UniqueName: \"kubernetes.io/projected/d30e2538-925e-4c64-8d47-7d0b7255f69d-kube-api-access-rq72k\") pod \"cert-manager-webhook-f4fb5df64-db8jl\" (UID: \"d30e2538-925e-4c64-8d47-7d0b7255f69d\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-db8jl" Jan 21 18:07:30 crc kubenswrapper[4808]: I0121 18:07:30.811677 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d30e2538-925e-4c64-8d47-7d0b7255f69d-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-db8jl\" (UID: \"d30e2538-925e-4c64-8d47-7d0b7255f69d\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-db8jl" Jan 21 18:07:30 crc kubenswrapper[4808]: I0121 18:07:30.838452 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq72k\" (UniqueName: \"kubernetes.io/projected/d30e2538-925e-4c64-8d47-7d0b7255f69d-kube-api-access-rq72k\") pod \"cert-manager-webhook-f4fb5df64-db8jl\" (UID: \"d30e2538-925e-4c64-8d47-7d0b7255f69d\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-db8jl" Jan 21 18:07:30 crc kubenswrapper[4808]: I0121 18:07:30.841909 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d30e2538-925e-4c64-8d47-7d0b7255f69d-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-db8jl\" (UID: \"d30e2538-925e-4c64-8d47-7d0b7255f69d\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-db8jl" Jan 21 18:07:30 crc kubenswrapper[4808]: I0121 18:07:30.950548 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-db8jl" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.177967 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-db8jl"] Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.291917 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.295181 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.301573 4808 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-fbwgb" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.301590 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-ca" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.301637 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-global-ca" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.301675 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-sys-config" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.310350 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.320227 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/3ba38831-b831-474c-8657-3fc4b335c8a3-builder-dockercfg-fbwgb-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.320312 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/3ba38831-b831-474c-8657-3fc4b335c8a3-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.320340 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3ba38831-b831-474c-8657-3fc4b335c8a3-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.320381 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.320405 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.320429 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/3ba38831-b831-474c-8657-3fc4b335c8a3-builder-dockercfg-fbwgb-push\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.320452 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snvkp\" (UniqueName: \"kubernetes.io/projected/3ba38831-b831-474c-8657-3fc4b335c8a3-kube-api-access-snvkp\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.320486 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3ba38831-b831-474c-8657-3fc4b335c8a3-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.320524 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/3ba38831-b831-474c-8657-3fc4b335c8a3-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.320547 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3ba38831-b831-474c-8657-3fc4b335c8a3-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.320566 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.320605 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.421955 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/3ba38831-b831-474c-8657-3fc4b335c8a3-builder-dockercfg-fbwgb-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.422013 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/3ba38831-b831-474c-8657-3fc4b335c8a3-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.422042 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3ba38831-b831-474c-8657-3fc4b335c8a3-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.422079 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.422097 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/3ba38831-b831-474c-8657-3fc4b335c8a3-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.422115 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.422153 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/3ba38831-b831-474c-8657-3fc4b335c8a3-builder-dockercfg-fbwgb-push\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.422178 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snvkp\" (UniqueName: \"kubernetes.io/projected/3ba38831-b831-474c-8657-3fc4b335c8a3-kube-api-access-snvkp\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.422210 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3ba38831-b831-474c-8657-3fc4b335c8a3-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.422252 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/3ba38831-b831-474c-8657-3fc4b335c8a3-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.422273 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3ba38831-b831-474c-8657-3fc4b335c8a3-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.422292 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.422328 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.422952 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.422967 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.423038 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3ba38831-b831-474c-8657-3fc4b335c8a3-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.423273 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.423786 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.423815 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/3ba38831-b831-474c-8657-3fc4b335c8a3-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.423949 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3ba38831-b831-474c-8657-3fc4b335c8a3-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.424472 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3ba38831-b831-474c-8657-3fc4b335c8a3-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.428109 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/3ba38831-b831-474c-8657-3fc4b335c8a3-builder-dockercfg-fbwgb-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.428129 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/3ba38831-b831-474c-8657-3fc4b335c8a3-builder-dockercfg-fbwgb-push\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.447377 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snvkp\" (UniqueName: \"kubernetes.io/projected/3ba38831-b831-474c-8657-3fc4b335c8a3-kube-api-access-snvkp\") pod \"service-telemetry-operator-1-build\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.616216 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.737136 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-db8jl" event={"ID":"d30e2538-925e-4c64-8d47-7d0b7255f69d","Type":"ContainerStarted","Data":"e265cdd235afb0069feb0a17210c41cb2e04720f34a9dc6f25a5690daa734dbd"} Jan 21 18:07:31 crc kubenswrapper[4808]: I0121 18:07:31.894522 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 21 18:07:32 crc kubenswrapper[4808]: I0121 18:07:32.746568 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"3ba38831-b831-474c-8657-3fc4b335c8a3","Type":"ContainerStarted","Data":"09f32d888fb1c0119028891cc578bf7f0486c4d9ceaf9689249138accf909c91"} Jan 21 18:07:34 crc kubenswrapper[4808]: I0121 18:07:34.512354 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-8jch4"] Jan 21 18:07:34 crc kubenswrapper[4808]: I0121 18:07:34.516450 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-8jch4" Jan 21 18:07:34 crc kubenswrapper[4808]: I0121 18:07:34.518907 4808 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-blgd5" Jan 21 18:07:34 crc kubenswrapper[4808]: I0121 18:07:34.522235 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-8jch4"] Jan 21 18:07:34 crc kubenswrapper[4808]: I0121 18:07:34.572202 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r74h\" (UniqueName: \"kubernetes.io/projected/71092cb7-08bc-4f0f-9696-838e2fc52d62-kube-api-access-4r74h\") pod \"cert-manager-cainjector-855d9ccff4-8jch4\" (UID: \"71092cb7-08bc-4f0f-9696-838e2fc52d62\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-8jch4" Jan 21 18:07:34 crc kubenswrapper[4808]: I0121 18:07:34.572317 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/71092cb7-08bc-4f0f-9696-838e2fc52d62-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-8jch4\" (UID: \"71092cb7-08bc-4f0f-9696-838e2fc52d62\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-8jch4" Jan 21 18:07:34 crc kubenswrapper[4808]: I0121 18:07:34.673913 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r74h\" (UniqueName: \"kubernetes.io/projected/71092cb7-08bc-4f0f-9696-838e2fc52d62-kube-api-access-4r74h\") pod \"cert-manager-cainjector-855d9ccff4-8jch4\" (UID: \"71092cb7-08bc-4f0f-9696-838e2fc52d62\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-8jch4" Jan 21 18:07:34 crc kubenswrapper[4808]: I0121 18:07:34.673986 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/71092cb7-08bc-4f0f-9696-838e2fc52d62-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-8jch4\" (UID: \"71092cb7-08bc-4f0f-9696-838e2fc52d62\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-8jch4" Jan 21 18:07:34 crc kubenswrapper[4808]: I0121 18:07:34.708748 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r74h\" (UniqueName: \"kubernetes.io/projected/71092cb7-08bc-4f0f-9696-838e2fc52d62-kube-api-access-4r74h\") pod \"cert-manager-cainjector-855d9ccff4-8jch4\" (UID: \"71092cb7-08bc-4f0f-9696-838e2fc52d62\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-8jch4" Jan 21 18:07:34 crc kubenswrapper[4808]: I0121 18:07:34.709609 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/71092cb7-08bc-4f0f-9696-838e2fc52d62-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-8jch4\" (UID: \"71092cb7-08bc-4f0f-9696-838e2fc52d62\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-8jch4" Jan 21 18:07:34 crc kubenswrapper[4808]: I0121 18:07:34.833832 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-8jch4" Jan 21 18:07:35 crc kubenswrapper[4808]: I0121 18:07:35.256228 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-8jch4"] Jan 21 18:07:35 crc kubenswrapper[4808]: W0121 18:07:35.278547 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71092cb7_08bc_4f0f_9696_838e2fc52d62.slice/crio-b160ef4556fff3695bdb2cf612d06ac80cca80130499963e8dacf3b2b8b114fa WatchSource:0}: Error finding container b160ef4556fff3695bdb2cf612d06ac80cca80130499963e8dacf3b2b8b114fa: Status 404 returned error can't find the container with id b160ef4556fff3695bdb2cf612d06ac80cca80130499963e8dacf3b2b8b114fa Jan 21 18:07:35 crc kubenswrapper[4808]: I0121 18:07:35.779359 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-8jch4" event={"ID":"71092cb7-08bc-4f0f-9696-838e2fc52d62","Type":"ContainerStarted","Data":"b160ef4556fff3695bdb2cf612d06ac80cca80130499963e8dacf3b2b8b114fa"} Jan 21 18:07:41 crc kubenswrapper[4808]: I0121 18:07:41.567180 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="a879c359-bceb-47c5-a3db-e41219bc698d" containerName="elasticsearch" probeResult="failure" output=< Jan 21 18:07:41 crc kubenswrapper[4808]: {"timestamp": "2026-01-21T18:07:41+00:00", "message": "readiness probe failed", "curl_rc": "7"} Jan 21 18:07:41 crc kubenswrapper[4808]: > Jan 21 18:07:41 crc kubenswrapper[4808]: I0121 18:07:41.674304 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.568763 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.571011 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.573003 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-global-ca" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.573421 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-sys-config" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.573550 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-ca" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.606762 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.606846 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.606918 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.606985 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-builder-dockercfg-fbwgb-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.607038 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.607082 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.607197 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.607224 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.607270 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.607289 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.607328 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bz7n\" (UniqueName: \"kubernetes.io/projected/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-kube-api-access-4bz7n\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.607355 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-builder-dockercfg-fbwgb-push\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.641117 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.709141 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.709210 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.709234 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.709285 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.709306 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.709329 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.709342 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bz7n\" (UniqueName: \"kubernetes.io/projected/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-kube-api-access-4bz7n\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.709491 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-builder-dockercfg-fbwgb-push\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.709535 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.709571 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.709624 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.709654 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-builder-dockercfg-fbwgb-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.709694 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.709746 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.710227 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.710226 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.710958 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.711460 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.713305 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.713389 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.713674 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.717870 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-builder-dockercfg-fbwgb-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.718222 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-builder-dockercfg-fbwgb-push\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.728894 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bz7n\" (UniqueName: \"kubernetes.io/projected/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-kube-api-access-4bz7n\") pod \"service-telemetry-operator-2-build\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:43 crc kubenswrapper[4808]: I0121 18:07:43.890670 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:46 crc kubenswrapper[4808]: I0121 18:07:46.819854 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/elasticsearch-es-default-0" Jan 21 18:07:49 crc kubenswrapper[4808]: E0121 18:07:49.050270 4808 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:29a0fa1c2f2a6cee62a0468a3883d16d491b4af29130dad6e3e2bb2948f274df" Jan 21 18:07:49 crc kubenswrapper[4808]: E0121 18:07:49.051518 4808 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cert-manager-webhook,Image:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:29a0fa1c2f2a6cee62a0468a3883d16d491b4af29130dad6e3e2bb2948f274df,Command:[/app/cmd/webhook/webhook],Args:[--dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=$(POD_NAMESPACE) --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.$(POD_NAMESPACE),cert-manager-webhook.$(POD_NAMESPACE).svc --secure-port=10250 --v=2],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:10250,Protocol:TCP,HostIP:,},ContainerPort{Name:healthcheck,HostPort:0,ContainerPort:6080,Protocol:TCP,HostIP:,},ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:9402,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:POD_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:bound-sa-token,ReadOnly:true,MountPath:/var/run/secrets/openshift/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rq72k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{1 0 healthcheck},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:60,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{1 0 healthcheck},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000690000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cert-manager-webhook-f4fb5df64-db8jl_cert-manager(d30e2538-925e-4c64-8d47-7d0b7255f69d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 21 18:07:49 crc kubenswrapper[4808]: E0121 18:07:49.053046 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="cert-manager/cert-manager-webhook-f4fb5df64-db8jl" podUID="d30e2538-925e-4c64-8d47-7d0b7255f69d" Jan 21 18:07:49 crc kubenswrapper[4808]: I0121 18:07:49.271832 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Jan 21 18:07:49 crc kubenswrapper[4808]: I0121 18:07:49.875955 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"a9c8f9d7-46a9-47a0-9f86-7b9183d90620","Type":"ContainerStarted","Data":"ab7a7a97467fb52d99147b49b21aad5200d0a4562285fd2abd150d37dc20e500"} Jan 21 18:07:49 crc kubenswrapper[4808]: I0121 18:07:49.876384 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"a9c8f9d7-46a9-47a0-9f86-7b9183d90620","Type":"ContainerStarted","Data":"c985f01a334036bbad2b498980eeb2d4f8704a03622449a77fc4f80ce196553b"} Jan 21 18:07:49 crc kubenswrapper[4808]: I0121 18:07:49.877426 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-8jch4" event={"ID":"71092cb7-08bc-4f0f-9696-838e2fc52d62","Type":"ContainerStarted","Data":"8de5e5f1fb6159c1ab994d00eab8aebbd1e77037e1bf0a3b2eff52488fd34595"} Jan 21 18:07:49 crc kubenswrapper[4808]: I0121 18:07:49.879411 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"3ba38831-b831-474c-8657-3fc4b335c8a3","Type":"ContainerStarted","Data":"1f138ca3fbee2f2338d97c32cff769bbf5d69829cc8af58302e2dc51c4c4cf05"} Jan 21 18:07:49 crc kubenswrapper[4808]: I0121 18:07:49.879562 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-1-build" podUID="3ba38831-b831-474c-8657-3fc4b335c8a3" containerName="manage-dockerfile" containerID="cri-o://1f138ca3fbee2f2338d97c32cff769bbf5d69829cc8af58302e2dc51c4c4cf05" gracePeriod=30 Jan 21 18:07:49 crc kubenswrapper[4808]: E0121 18:07:49.941669 4808 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=7674227450993299356, SKID=, AKID=DF:1F:FD:7A:27:6A:84:51:E7:A3:2B:77:06:EA:03:AB:6E:1F:D3:D1 failed: x509: certificate signed by unknown authority" Jan 21 18:07:49 crc kubenswrapper[4808]: I0121 18:07:49.971711 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-8jch4" podStartSLOduration=1.940270369 podStartE2EDuration="15.971689928s" podCreationTimestamp="2026-01-21 18:07:34 +0000 UTC" firstStartedPulling="2026-01-21 18:07:35.283594025 +0000 UTC m=+690.014104910" lastFinishedPulling="2026-01-21 18:07:49.315013584 +0000 UTC m=+704.045524469" observedRunningTime="2026-01-21 18:07:49.965769092 +0000 UTC m=+704.696279997" watchObservedRunningTime="2026-01-21 18:07:49.971689928 +0000 UTC m=+704.702200823" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.386219 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_3ba38831-b831-474c-8657-3fc4b335c8a3/manage-dockerfile/0.log" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.386557 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.547054 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/3ba38831-b831-474c-8657-3fc4b335c8a3-builder-dockercfg-fbwgb-push\") pod \"3ba38831-b831-474c-8657-3fc4b335c8a3\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.547107 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3ba38831-b831-474c-8657-3fc4b335c8a3-build-ca-bundles\") pod \"3ba38831-b831-474c-8657-3fc4b335c8a3\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.547144 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/3ba38831-b831-474c-8657-3fc4b335c8a3-builder-dockercfg-fbwgb-pull\") pod \"3ba38831-b831-474c-8657-3fc4b335c8a3\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.547217 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3ba38831-b831-474c-8657-3fc4b335c8a3-build-proxy-ca-bundles\") pod \"3ba38831-b831-474c-8657-3fc4b335c8a3\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.547263 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-container-storage-root\") pod \"3ba38831-b831-474c-8657-3fc4b335c8a3\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.547310 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/3ba38831-b831-474c-8657-3fc4b335c8a3-build-system-configs\") pod \"3ba38831-b831-474c-8657-3fc4b335c8a3\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.547339 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/3ba38831-b831-474c-8657-3fc4b335c8a3-buildcachedir\") pod \"3ba38831-b831-474c-8657-3fc4b335c8a3\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.547374 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-build-blob-cache\") pod \"3ba38831-b831-474c-8657-3fc4b335c8a3\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.547399 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-buildworkdir\") pod \"3ba38831-b831-474c-8657-3fc4b335c8a3\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.547465 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3ba38831-b831-474c-8657-3fc4b335c8a3-node-pullsecrets\") pod \"3ba38831-b831-474c-8657-3fc4b335c8a3\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.547499 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snvkp\" (UniqueName: \"kubernetes.io/projected/3ba38831-b831-474c-8657-3fc4b335c8a3-kube-api-access-snvkp\") pod \"3ba38831-b831-474c-8657-3fc4b335c8a3\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.547545 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-container-storage-run\") pod \"3ba38831-b831-474c-8657-3fc4b335c8a3\" (UID: \"3ba38831-b831-474c-8657-3fc4b335c8a3\") " Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.548126 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "3ba38831-b831-474c-8657-3fc4b335c8a3" (UID: "3ba38831-b831-474c-8657-3fc4b335c8a3"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.548329 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3ba38831-b831-474c-8657-3fc4b335c8a3-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "3ba38831-b831-474c-8657-3fc4b335c8a3" (UID: "3ba38831-b831-474c-8657-3fc4b335c8a3"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.548619 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "3ba38831-b831-474c-8657-3fc4b335c8a3" (UID: "3ba38831-b831-474c-8657-3fc4b335c8a3"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.548753 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "3ba38831-b831-474c-8657-3fc4b335c8a3" (UID: "3ba38831-b831-474c-8657-3fc4b335c8a3"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.548798 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "3ba38831-b831-474c-8657-3fc4b335c8a3" (UID: "3ba38831-b831-474c-8657-3fc4b335c8a3"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.548978 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ba38831-b831-474c-8657-3fc4b335c8a3-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "3ba38831-b831-474c-8657-3fc4b335c8a3" (UID: "3ba38831-b831-474c-8657-3fc4b335c8a3"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.549012 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ba38831-b831-474c-8657-3fc4b335c8a3-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "3ba38831-b831-474c-8657-3fc4b335c8a3" (UID: "3ba38831-b831-474c-8657-3fc4b335c8a3"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.549010 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3ba38831-b831-474c-8657-3fc4b335c8a3-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "3ba38831-b831-474c-8657-3fc4b335c8a3" (UID: "3ba38831-b831-474c-8657-3fc4b335c8a3"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.549301 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ba38831-b831-474c-8657-3fc4b335c8a3-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "3ba38831-b831-474c-8657-3fc4b335c8a3" (UID: "3ba38831-b831-474c-8657-3fc4b335c8a3"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.553330 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ba38831-b831-474c-8657-3fc4b335c8a3-builder-dockercfg-fbwgb-push" (OuterVolumeSpecName: "builder-dockercfg-fbwgb-push") pod "3ba38831-b831-474c-8657-3fc4b335c8a3" (UID: "3ba38831-b831-474c-8657-3fc4b335c8a3"). InnerVolumeSpecName "builder-dockercfg-fbwgb-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.553540 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ba38831-b831-474c-8657-3fc4b335c8a3-builder-dockercfg-fbwgb-pull" (OuterVolumeSpecName: "builder-dockercfg-fbwgb-pull") pod "3ba38831-b831-474c-8657-3fc4b335c8a3" (UID: "3ba38831-b831-474c-8657-3fc4b335c8a3"). InnerVolumeSpecName "builder-dockercfg-fbwgb-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.555359 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ba38831-b831-474c-8657-3fc4b335c8a3-kube-api-access-snvkp" (OuterVolumeSpecName: "kube-api-access-snvkp") pod "3ba38831-b831-474c-8657-3fc4b335c8a3" (UID: "3ba38831-b831-474c-8657-3fc4b335c8a3"). InnerVolumeSpecName "kube-api-access-snvkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.648789 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snvkp\" (UniqueName: \"kubernetes.io/projected/3ba38831-b831-474c-8657-3fc4b335c8a3-kube-api-access-snvkp\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.648841 4808 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.648854 4808 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/3ba38831-b831-474c-8657-3fc4b335c8a3-builder-dockercfg-fbwgb-push\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.648869 4808 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3ba38831-b831-474c-8657-3fc4b335c8a3-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.648881 4808 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/3ba38831-b831-474c-8657-3fc4b335c8a3-builder-dockercfg-fbwgb-pull\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.648890 4808 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.648897 4808 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3ba38831-b831-474c-8657-3fc4b335c8a3-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.648905 4808 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/3ba38831-b831-474c-8657-3fc4b335c8a3-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.648914 4808 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/3ba38831-b831-474c-8657-3fc4b335c8a3-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.648923 4808 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.648931 4808 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/3ba38831-b831-474c-8657-3fc4b335c8a3-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.648940 4808 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3ba38831-b831-474c-8657-3fc4b335c8a3-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.745980 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-mvjvw"] Jan 21 18:07:50 crc kubenswrapper[4808]: E0121 18:07:50.746348 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ba38831-b831-474c-8657-3fc4b335c8a3" containerName="manage-dockerfile" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.746369 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ba38831-b831-474c-8657-3fc4b335c8a3" containerName="manage-dockerfile" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.746490 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ba38831-b831-474c-8657-3fc4b335c8a3" containerName="manage-dockerfile" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.746906 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-mvjvw" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.749779 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwzld\" (UniqueName: \"kubernetes.io/projected/045426c7-6605-4623-93c3-1575773a5ecc-kube-api-access-nwzld\") pod \"cert-manager-86cb77c54b-mvjvw\" (UID: \"045426c7-6605-4623-93c3-1575773a5ecc\") " pod="cert-manager/cert-manager-86cb77c54b-mvjvw" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.749876 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/045426c7-6605-4623-93c3-1575773a5ecc-bound-sa-token\") pod \"cert-manager-86cb77c54b-mvjvw\" (UID: \"045426c7-6605-4623-93c3-1575773a5ecc\") " pod="cert-manager/cert-manager-86cb77c54b-mvjvw" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.751930 4808 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-9jmvn" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.759233 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-mvjvw"] Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.850631 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/045426c7-6605-4623-93c3-1575773a5ecc-bound-sa-token\") pod \"cert-manager-86cb77c54b-mvjvw\" (UID: \"045426c7-6605-4623-93c3-1575773a5ecc\") " pod="cert-manager/cert-manager-86cb77c54b-mvjvw" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.851082 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwzld\" (UniqueName: \"kubernetes.io/projected/045426c7-6605-4623-93c3-1575773a5ecc-kube-api-access-nwzld\") pod \"cert-manager-86cb77c54b-mvjvw\" (UID: \"045426c7-6605-4623-93c3-1575773a5ecc\") " pod="cert-manager/cert-manager-86cb77c54b-mvjvw" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.866125 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwzld\" (UniqueName: \"kubernetes.io/projected/045426c7-6605-4623-93c3-1575773a5ecc-kube-api-access-nwzld\") pod \"cert-manager-86cb77c54b-mvjvw\" (UID: \"045426c7-6605-4623-93c3-1575773a5ecc\") " pod="cert-manager/cert-manager-86cb77c54b-mvjvw" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.868793 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/045426c7-6605-4623-93c3-1575773a5ecc-bound-sa-token\") pod \"cert-manager-86cb77c54b-mvjvw\" (UID: \"045426c7-6605-4623-93c3-1575773a5ecc\") " pod="cert-manager/cert-manager-86cb77c54b-mvjvw" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.887285 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_3ba38831-b831-474c-8657-3fc4b335c8a3/manage-dockerfile/0.log" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.887339 4808 generic.go:334] "Generic (PLEG): container finished" podID="3ba38831-b831-474c-8657-3fc4b335c8a3" containerID="1f138ca3fbee2f2338d97c32cff769bbf5d69829cc8af58302e2dc51c4c4cf05" exitCode=1 Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.887410 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"3ba38831-b831-474c-8657-3fc4b335c8a3","Type":"ContainerDied","Data":"1f138ca3fbee2f2338d97c32cff769bbf5d69829cc8af58302e2dc51c4c4cf05"} Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.887495 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"3ba38831-b831-474c-8657-3fc4b335c8a3","Type":"ContainerDied","Data":"09f32d888fb1c0119028891cc578bf7f0486c4d9ceaf9689249138accf909c91"} Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.887560 4808 scope.go:117] "RemoveContainer" containerID="1f138ca3fbee2f2338d97c32cff769bbf5d69829cc8af58302e2dc51c4c4cf05" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.887815 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.889754 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-db8jl" event={"ID":"d30e2538-925e-4c64-8d47-7d0b7255f69d","Type":"ContainerStarted","Data":"23c58421462f70d4d7734b50de5c64ea51edd7761053fe187973676366985861"} Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.890613 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-db8jl" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.917256 4808 scope.go:117] "RemoveContainer" containerID="1f138ca3fbee2f2338d97c32cff769bbf5d69829cc8af58302e2dc51c4c4cf05" Jan 21 18:07:50 crc kubenswrapper[4808]: E0121 18:07:50.917846 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f138ca3fbee2f2338d97c32cff769bbf5d69829cc8af58302e2dc51c4c4cf05\": container with ID starting with 1f138ca3fbee2f2338d97c32cff769bbf5d69829cc8af58302e2dc51c4c4cf05 not found: ID does not exist" containerID="1f138ca3fbee2f2338d97c32cff769bbf5d69829cc8af58302e2dc51c4c4cf05" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.917889 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f138ca3fbee2f2338d97c32cff769bbf5d69829cc8af58302e2dc51c4c4cf05"} err="failed to get container status \"1f138ca3fbee2f2338d97c32cff769bbf5d69829cc8af58302e2dc51c4c4cf05\": rpc error: code = NotFound desc = could not find container \"1f138ca3fbee2f2338d97c32cff769bbf5d69829cc8af58302e2dc51c4c4cf05\": container with ID starting with 1f138ca3fbee2f2338d97c32cff769bbf5d69829cc8af58302e2dc51c4c4cf05 not found: ID does not exist" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.945163 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-db8jl" podStartSLOduration=-9223372015.909634 podStartE2EDuration="20.94514289s" podCreationTimestamp="2026-01-21 18:07:30 +0000 UTC" firstStartedPulling="2026-01-21 18:07:31.19040565 +0000 UTC m=+685.920916535" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 18:07:50.926118832 +0000 UTC m=+705.656629717" watchObservedRunningTime="2026-01-21 18:07:50.94514289 +0000 UTC m=+705.675653785" Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.954311 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.971172 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 21 18:07:50 crc kubenswrapper[4808]: I0121 18:07:50.988073 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Jan 21 18:07:51 crc kubenswrapper[4808]: I0121 18:07:51.061360 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-mvjvw" Jan 21 18:07:51 crc kubenswrapper[4808]: I0121 18:07:51.293995 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-mvjvw"] Jan 21 18:07:51 crc kubenswrapper[4808]: W0121 18:07:51.301329 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod045426c7_6605_4623_93c3_1575773a5ecc.slice/crio-a4d227e4f52b084f5bfc4440baffeec06bcca922ba848d6c1fb29e32d8a9a406 WatchSource:0}: Error finding container a4d227e4f52b084f5bfc4440baffeec06bcca922ba848d6c1fb29e32d8a9a406: Status 404 returned error can't find the container with id a4d227e4f52b084f5bfc4440baffeec06bcca922ba848d6c1fb29e32d8a9a406 Jan 21 18:07:51 crc kubenswrapper[4808]: I0121 18:07:51.533391 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ba38831-b831-474c-8657-3fc4b335c8a3" path="/var/lib/kubelet/pods/3ba38831-b831-474c-8657-3fc4b335c8a3/volumes" Jan 21 18:07:51 crc kubenswrapper[4808]: I0121 18:07:51.901807 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-mvjvw" event={"ID":"045426c7-6605-4623-93c3-1575773a5ecc","Type":"ContainerStarted","Data":"022a7538ace24684bb97daa9beaa247e76539dc3a10aab69467f00e778ab7f2a"} Jan 21 18:07:51 crc kubenswrapper[4808]: I0121 18:07:51.901851 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-mvjvw" event={"ID":"045426c7-6605-4623-93c3-1575773a5ecc","Type":"ContainerStarted","Data":"a4d227e4f52b084f5bfc4440baffeec06bcca922ba848d6c1fb29e32d8a9a406"} Jan 21 18:07:51 crc kubenswrapper[4808]: I0121 18:07:51.905564 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-2-build" podUID="a9c8f9d7-46a9-47a0-9f86-7b9183d90620" containerName="git-clone" containerID="cri-o://ab7a7a97467fb52d99147b49b21aad5200d0a4562285fd2abd150d37dc20e500" gracePeriod=30 Jan 21 18:07:51 crc kubenswrapper[4808]: I0121 18:07:51.918544 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-mvjvw" podStartSLOduration=1.918517149 podStartE2EDuration="1.918517149s" podCreationTimestamp="2026-01-21 18:07:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 18:07:51.916876059 +0000 UTC m=+706.647386954" watchObservedRunningTime="2026-01-21 18:07:51.918517149 +0000 UTC m=+706.649028034" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.286692 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_a9c8f9d7-46a9-47a0-9f86-7b9183d90620/git-clone/0.log" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.286969 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.475559 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-node-pullsecrets\") pod \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.475634 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-blob-cache\") pod \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.475699 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bz7n\" (UniqueName: \"kubernetes.io/projected/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-kube-api-access-4bz7n\") pod \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.475728 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-container-storage-run\") pod \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.475759 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-container-storage-root\") pod \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.475816 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-buildworkdir\") pod \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.475840 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-buildcachedir\") pod \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.475885 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-proxy-ca-bundles\") pod \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.475920 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-ca-bundles\") pod \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.475952 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-system-configs\") pod \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.475989 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-builder-dockercfg-fbwgb-pull\") pod \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.476029 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-builder-dockercfg-fbwgb-push\") pod \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\" (UID: \"a9c8f9d7-46a9-47a0-9f86-7b9183d90620\") " Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.476306 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "a9c8f9d7-46a9-47a0-9f86-7b9183d90620" (UID: "a9c8f9d7-46a9-47a0-9f86-7b9183d90620"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.476368 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "a9c8f9d7-46a9-47a0-9f86-7b9183d90620" (UID: "a9c8f9d7-46a9-47a0-9f86-7b9183d90620"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.476751 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "a9c8f9d7-46a9-47a0-9f86-7b9183d90620" (UID: "a9c8f9d7-46a9-47a0-9f86-7b9183d90620"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.476961 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "a9c8f9d7-46a9-47a0-9f86-7b9183d90620" (UID: "a9c8f9d7-46a9-47a0-9f86-7b9183d90620"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.477113 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "a9c8f9d7-46a9-47a0-9f86-7b9183d90620" (UID: "a9c8f9d7-46a9-47a0-9f86-7b9183d90620"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.477317 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "a9c8f9d7-46a9-47a0-9f86-7b9183d90620" (UID: "a9c8f9d7-46a9-47a0-9f86-7b9183d90620"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.477342 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "a9c8f9d7-46a9-47a0-9f86-7b9183d90620" (UID: "a9c8f9d7-46a9-47a0-9f86-7b9183d90620"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.477539 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "a9c8f9d7-46a9-47a0-9f86-7b9183d90620" (UID: "a9c8f9d7-46a9-47a0-9f86-7b9183d90620"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.477896 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "a9c8f9d7-46a9-47a0-9f86-7b9183d90620" (UID: "a9c8f9d7-46a9-47a0-9f86-7b9183d90620"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.482456 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-builder-dockercfg-fbwgb-pull" (OuterVolumeSpecName: "builder-dockercfg-fbwgb-pull") pod "a9c8f9d7-46a9-47a0-9f86-7b9183d90620" (UID: "a9c8f9d7-46a9-47a0-9f86-7b9183d90620"). InnerVolumeSpecName "builder-dockercfg-fbwgb-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.482496 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-builder-dockercfg-fbwgb-push" (OuterVolumeSpecName: "builder-dockercfg-fbwgb-push") pod "a9c8f9d7-46a9-47a0-9f86-7b9183d90620" (UID: "a9c8f9d7-46a9-47a0-9f86-7b9183d90620"). InnerVolumeSpecName "builder-dockercfg-fbwgb-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.484998 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-kube-api-access-4bz7n" (OuterVolumeSpecName: "kube-api-access-4bz7n") pod "a9c8f9d7-46a9-47a0-9f86-7b9183d90620" (UID: "a9c8f9d7-46a9-47a0-9f86-7b9183d90620"). InnerVolumeSpecName "kube-api-access-4bz7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.577430 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bz7n\" (UniqueName: \"kubernetes.io/projected/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-kube-api-access-4bz7n\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.577465 4808 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.577477 4808 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.577486 4808 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.577494 4808 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.577502 4808 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.577510 4808 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.577518 4808 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.577527 4808 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-builder-dockercfg-fbwgb-pull\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.577537 4808 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-builder-dockercfg-fbwgb-push\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.577545 4808 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.577555 4808 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a9c8f9d7-46a9-47a0-9f86-7b9183d90620-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.913145 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_a9c8f9d7-46a9-47a0-9f86-7b9183d90620/git-clone/0.log" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.913188 4808 generic.go:334] "Generic (PLEG): container finished" podID="a9c8f9d7-46a9-47a0-9f86-7b9183d90620" containerID="ab7a7a97467fb52d99147b49b21aad5200d0a4562285fd2abd150d37dc20e500" exitCode=1 Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.913885 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.916400 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"a9c8f9d7-46a9-47a0-9f86-7b9183d90620","Type":"ContainerDied","Data":"ab7a7a97467fb52d99147b49b21aad5200d0a4562285fd2abd150d37dc20e500"} Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.916474 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"a9c8f9d7-46a9-47a0-9f86-7b9183d90620","Type":"ContainerDied","Data":"c985f01a334036bbad2b498980eeb2d4f8704a03622449a77fc4f80ce196553b"} Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.916497 4808 scope.go:117] "RemoveContainer" containerID="ab7a7a97467fb52d99147b49b21aad5200d0a4562285fd2abd150d37dc20e500" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.940016 4808 scope.go:117] "RemoveContainer" containerID="ab7a7a97467fb52d99147b49b21aad5200d0a4562285fd2abd150d37dc20e500" Jan 21 18:07:52 crc kubenswrapper[4808]: E0121 18:07:52.942631 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab7a7a97467fb52d99147b49b21aad5200d0a4562285fd2abd150d37dc20e500\": container with ID starting with ab7a7a97467fb52d99147b49b21aad5200d0a4562285fd2abd150d37dc20e500 not found: ID does not exist" containerID="ab7a7a97467fb52d99147b49b21aad5200d0a4562285fd2abd150d37dc20e500" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.942676 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab7a7a97467fb52d99147b49b21aad5200d0a4562285fd2abd150d37dc20e500"} err="failed to get container status \"ab7a7a97467fb52d99147b49b21aad5200d0a4562285fd2abd150d37dc20e500\": rpc error: code = NotFound desc = could not find container \"ab7a7a97467fb52d99147b49b21aad5200d0a4562285fd2abd150d37dc20e500\": container with ID starting with ab7a7a97467fb52d99147b49b21aad5200d0a4562285fd2abd150d37dc20e500 not found: ID does not exist" Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.950104 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Jan 21 18:07:52 crc kubenswrapper[4808]: I0121 18:07:52.958063 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Jan 21 18:07:53 crc kubenswrapper[4808]: I0121 18:07:53.529581 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9c8f9d7-46a9-47a0-9f86-7b9183d90620" path="/var/lib/kubelet/pods/a9c8f9d7-46a9-47a0-9f86-7b9183d90620/volumes" Jan 21 18:07:55 crc kubenswrapper[4808]: I0121 18:07:55.954469 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-db8jl" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.518696 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Jan 21 18:08:02 crc kubenswrapper[4808]: E0121 18:08:02.519774 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9c8f9d7-46a9-47a0-9f86-7b9183d90620" containerName="git-clone" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.519796 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9c8f9d7-46a9-47a0-9f86-7b9183d90620" containerName="git-clone" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.520007 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9c8f9d7-46a9-47a0-9f86-7b9183d90620" containerName="git-clone" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.521475 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.523065 4808 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-fbwgb" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.524118 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-3-ca" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.524165 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-3-sys-config" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.524205 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-3-global-ca" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.548076 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.620701 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7d4d88b4-a127-465d-b131-e37b28c46140-buildcachedir\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.620757 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7d4d88b4-a127-465d-b131-e37b28c46140-node-pullsecrets\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.620788 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/7d4d88b4-a127-465d-b131-e37b28c46140-builder-dockercfg-fbwgb-push\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.620808 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jhhg\" (UniqueName: \"kubernetes.io/projected/7d4d88b4-a127-465d-b131-e37b28c46140-kube-api-access-5jhhg\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.620836 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7d4d88b4-a127-465d-b131-e37b28c46140-build-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.620876 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-container-storage-root\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.620933 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/7d4d88b4-a127-465d-b131-e37b28c46140-builder-dockercfg-fbwgb-pull\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.620963 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-buildworkdir\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.620999 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-container-storage-run\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.621030 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7d4d88b4-a127-465d-b131-e37b28c46140-build-proxy-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.621081 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7d4d88b4-a127-465d-b131-e37b28c46140-build-system-configs\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.621106 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-build-blob-cache\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.723046 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7d4d88b4-a127-465d-b131-e37b28c46140-build-system-configs\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.723123 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-build-blob-cache\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.723224 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7d4d88b4-a127-465d-b131-e37b28c46140-buildcachedir\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.723312 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7d4d88b4-a127-465d-b131-e37b28c46140-node-pullsecrets\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.723345 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/7d4d88b4-a127-465d-b131-e37b28c46140-builder-dockercfg-fbwgb-push\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.723371 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jhhg\" (UniqueName: \"kubernetes.io/projected/7d4d88b4-a127-465d-b131-e37b28c46140-kube-api-access-5jhhg\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.723406 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7d4d88b4-a127-465d-b131-e37b28c46140-build-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.723415 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7d4d88b4-a127-465d-b131-e37b28c46140-buildcachedir\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.723458 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-container-storage-root\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.723433 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7d4d88b4-a127-465d-b131-e37b28c46140-node-pullsecrets\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.723576 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/7d4d88b4-a127-465d-b131-e37b28c46140-builder-dockercfg-fbwgb-pull\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.723650 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-build-blob-cache\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.723654 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-buildworkdir\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.723747 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-container-storage-run\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.723796 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7d4d88b4-a127-465d-b131-e37b28c46140-build-proxy-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.724505 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7d4d88b4-a127-465d-b131-e37b28c46140-build-system-configs\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.724613 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7d4d88b4-a127-465d-b131-e37b28c46140-build-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.724613 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7d4d88b4-a127-465d-b131-e37b28c46140-build-proxy-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.724856 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-container-storage-root\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.725022 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-container-storage-run\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.725162 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-buildworkdir\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.728859 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/7d4d88b4-a127-465d-b131-e37b28c46140-builder-dockercfg-fbwgb-pull\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.733583 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/7d4d88b4-a127-465d-b131-e37b28c46140-builder-dockercfg-fbwgb-push\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.747649 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jhhg\" (UniqueName: \"kubernetes.io/projected/7d4d88b4-a127-465d-b131-e37b28c46140-kube-api-access-5jhhg\") pod \"service-telemetry-operator-3-build\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:02 crc kubenswrapper[4808]: I0121 18:08:02.843949 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:03 crc kubenswrapper[4808]: I0121 18:08:03.254179 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Jan 21 18:08:03 crc kubenswrapper[4808]: I0121 18:08:03.987578 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-3-build" event={"ID":"7d4d88b4-a127-465d-b131-e37b28c46140","Type":"ContainerStarted","Data":"687841d572f81e2ca13e21701a9a7e4d3101efb81420c3c5c6c90b4fb9e04142"} Jan 21 18:08:06 crc kubenswrapper[4808]: I0121 18:08:06.004063 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-3-build" event={"ID":"7d4d88b4-a127-465d-b131-e37b28c46140","Type":"ContainerStarted","Data":"8b46d6f18580bc0cbd53d1d708bab0b1c771500727ffba7a0b372723febe0f5b"} Jan 21 18:08:06 crc kubenswrapper[4808]: E0121 18:08:06.065631 4808 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=7674227450993299356, SKID=, AKID=DF:1F:FD:7A:27:6A:84:51:E7:A3:2B:77:06:EA:03:AB:6E:1F:D3:D1 failed: x509: certificate signed by unknown authority" Jan 21 18:08:07 crc kubenswrapper[4808]: I0121 18:08:07.098399 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.016815 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-3-build" podUID="7d4d88b4-a127-465d-b131-e37b28c46140" containerName="git-clone" containerID="cri-o://8b46d6f18580bc0cbd53d1d708bab0b1c771500727ffba7a0b372723febe0f5b" gracePeriod=30 Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.395370 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-3-build_7d4d88b4-a127-465d-b131-e37b28c46140/git-clone/0.log" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.395718 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.399936 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/7d4d88b4-a127-465d-b131-e37b28c46140-builder-dockercfg-fbwgb-pull\") pod \"7d4d88b4-a127-465d-b131-e37b28c46140\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.399995 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/7d4d88b4-a127-465d-b131-e37b28c46140-builder-dockercfg-fbwgb-push\") pod \"7d4d88b4-a127-465d-b131-e37b28c46140\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.400062 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-container-storage-root\") pod \"7d4d88b4-a127-465d-b131-e37b28c46140\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.400084 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-build-blob-cache\") pod \"7d4d88b4-a127-465d-b131-e37b28c46140\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.400498 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "7d4d88b4-a127-465d-b131-e37b28c46140" (UID: "7d4d88b4-a127-465d-b131-e37b28c46140"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.400738 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "7d4d88b4-a127-465d-b131-e37b28c46140" (UID: "7d4d88b4-a127-465d-b131-e37b28c46140"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.401303 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7d4d88b4-a127-465d-b131-e37b28c46140-build-proxy-ca-bundles\") pod \"7d4d88b4-a127-465d-b131-e37b28c46140\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.401368 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7d4d88b4-a127-465d-b131-e37b28c46140-build-system-configs\") pod \"7d4d88b4-a127-465d-b131-e37b28c46140\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.401404 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7d4d88b4-a127-465d-b131-e37b28c46140-buildcachedir\") pod \"7d4d88b4-a127-465d-b131-e37b28c46140\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.401458 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-buildworkdir\") pod \"7d4d88b4-a127-465d-b131-e37b28c46140\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.401840 4808 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.401858 4808 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.402083 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d4d88b4-a127-465d-b131-e37b28c46140-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "7d4d88b4-a127-465d-b131-e37b28c46140" (UID: "7d4d88b4-a127-465d-b131-e37b28c46140"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.402092 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d4d88b4-a127-465d-b131-e37b28c46140-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "7d4d88b4-a127-465d-b131-e37b28c46140" (UID: "7d4d88b4-a127-465d-b131-e37b28c46140"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.402209 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "7d4d88b4-a127-465d-b131-e37b28c46140" (UID: "7d4d88b4-a127-465d-b131-e37b28c46140"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.402330 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d4d88b4-a127-465d-b131-e37b28c46140-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "7d4d88b4-a127-465d-b131-e37b28c46140" (UID: "7d4d88b4-a127-465d-b131-e37b28c46140"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.405470 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d4d88b4-a127-465d-b131-e37b28c46140-builder-dockercfg-fbwgb-pull" (OuterVolumeSpecName: "builder-dockercfg-fbwgb-pull") pod "7d4d88b4-a127-465d-b131-e37b28c46140" (UID: "7d4d88b4-a127-465d-b131-e37b28c46140"). InnerVolumeSpecName "builder-dockercfg-fbwgb-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.405489 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d4d88b4-a127-465d-b131-e37b28c46140-builder-dockercfg-fbwgb-push" (OuterVolumeSpecName: "builder-dockercfg-fbwgb-push") pod "7d4d88b4-a127-465d-b131-e37b28c46140" (UID: "7d4d88b4-a127-465d-b131-e37b28c46140"). InnerVolumeSpecName "builder-dockercfg-fbwgb-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.502681 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7d4d88b4-a127-465d-b131-e37b28c46140-build-ca-bundles\") pod \"7d4d88b4-a127-465d-b131-e37b28c46140\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.502905 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jhhg\" (UniqueName: \"kubernetes.io/projected/7d4d88b4-a127-465d-b131-e37b28c46140-kube-api-access-5jhhg\") pod \"7d4d88b4-a127-465d-b131-e37b28c46140\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.502943 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7d4d88b4-a127-465d-b131-e37b28c46140-node-pullsecrets\") pod \"7d4d88b4-a127-465d-b131-e37b28c46140\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.502998 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-container-storage-run\") pod \"7d4d88b4-a127-465d-b131-e37b28c46140\" (UID: \"7d4d88b4-a127-465d-b131-e37b28c46140\") " Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.503040 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d4d88b4-a127-465d-b131-e37b28c46140-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "7d4d88b4-a127-465d-b131-e37b28c46140" (UID: "7d4d88b4-a127-465d-b131-e37b28c46140"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.503332 4808 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/7d4d88b4-a127-465d-b131-e37b28c46140-builder-dockercfg-fbwgb-pull\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.503356 4808 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/7d4d88b4-a127-465d-b131-e37b28c46140-builder-dockercfg-fbwgb-push\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.503369 4808 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7d4d88b4-a127-465d-b131-e37b28c46140-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.503381 4808 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7d4d88b4-a127-465d-b131-e37b28c46140-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.503393 4808 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7d4d88b4-a127-465d-b131-e37b28c46140-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.503403 4808 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.503414 4808 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7d4d88b4-a127-465d-b131-e37b28c46140-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.503338 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d4d88b4-a127-465d-b131-e37b28c46140-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "7d4d88b4-a127-465d-b131-e37b28c46140" (UID: "7d4d88b4-a127-465d-b131-e37b28c46140"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.503419 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "7d4d88b4-a127-465d-b131-e37b28c46140" (UID: "7d4d88b4-a127-465d-b131-e37b28c46140"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.506061 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d4d88b4-a127-465d-b131-e37b28c46140-kube-api-access-5jhhg" (OuterVolumeSpecName: "kube-api-access-5jhhg") pod "7d4d88b4-a127-465d-b131-e37b28c46140" (UID: "7d4d88b4-a127-465d-b131-e37b28c46140"). InnerVolumeSpecName "kube-api-access-5jhhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.604638 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jhhg\" (UniqueName: \"kubernetes.io/projected/7d4d88b4-a127-465d-b131-e37b28c46140-kube-api-access-5jhhg\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.604683 4808 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7d4d88b4-a127-465d-b131-e37b28c46140-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:08 crc kubenswrapper[4808]: I0121 18:08:08.604696 4808 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7d4d88b4-a127-465d-b131-e37b28c46140-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:09 crc kubenswrapper[4808]: I0121 18:08:09.025313 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-3-build_7d4d88b4-a127-465d-b131-e37b28c46140/git-clone/0.log" Jan 21 18:08:09 crc kubenswrapper[4808]: I0121 18:08:09.025416 4808 generic.go:334] "Generic (PLEG): container finished" podID="7d4d88b4-a127-465d-b131-e37b28c46140" containerID="8b46d6f18580bc0cbd53d1d708bab0b1c771500727ffba7a0b372723febe0f5b" exitCode=1 Jan 21 18:08:09 crc kubenswrapper[4808]: I0121 18:08:09.025471 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-3-build" event={"ID":"7d4d88b4-a127-465d-b131-e37b28c46140","Type":"ContainerDied","Data":"8b46d6f18580bc0cbd53d1d708bab0b1c771500727ffba7a0b372723febe0f5b"} Jan 21 18:08:09 crc kubenswrapper[4808]: I0121 18:08:09.025496 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-3-build" Jan 21 18:08:09 crc kubenswrapper[4808]: I0121 18:08:09.025513 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-3-build" event={"ID":"7d4d88b4-a127-465d-b131-e37b28c46140","Type":"ContainerDied","Data":"687841d572f81e2ca13e21701a9a7e4d3101efb81420c3c5c6c90b4fb9e04142"} Jan 21 18:08:09 crc kubenswrapper[4808]: I0121 18:08:09.025555 4808 scope.go:117] "RemoveContainer" containerID="8b46d6f18580bc0cbd53d1d708bab0b1c771500727ffba7a0b372723febe0f5b" Jan 21 18:08:09 crc kubenswrapper[4808]: I0121 18:08:09.043759 4808 scope.go:117] "RemoveContainer" containerID="8b46d6f18580bc0cbd53d1d708bab0b1c771500727ffba7a0b372723febe0f5b" Jan 21 18:08:09 crc kubenswrapper[4808]: E0121 18:08:09.044318 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b46d6f18580bc0cbd53d1d708bab0b1c771500727ffba7a0b372723febe0f5b\": container with ID starting with 8b46d6f18580bc0cbd53d1d708bab0b1c771500727ffba7a0b372723febe0f5b not found: ID does not exist" containerID="8b46d6f18580bc0cbd53d1d708bab0b1c771500727ffba7a0b372723febe0f5b" Jan 21 18:08:09 crc kubenswrapper[4808]: I0121 18:08:09.044379 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b46d6f18580bc0cbd53d1d708bab0b1c771500727ffba7a0b372723febe0f5b"} err="failed to get container status \"8b46d6f18580bc0cbd53d1d708bab0b1c771500727ffba7a0b372723febe0f5b\": rpc error: code = NotFound desc = could not find container \"8b46d6f18580bc0cbd53d1d708bab0b1c771500727ffba7a0b372723febe0f5b\": container with ID starting with 8b46d6f18580bc0cbd53d1d708bab0b1c771500727ffba7a0b372723febe0f5b not found: ID does not exist" Jan 21 18:08:09 crc kubenswrapper[4808]: I0121 18:08:09.070199 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Jan 21 18:08:09 crc kubenswrapper[4808]: I0121 18:08:09.082511 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Jan 21 18:08:09 crc kubenswrapper[4808]: I0121 18:08:09.530707 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d4d88b4-a127-465d-b131-e37b28c46140" path="/var/lib/kubelet/pods/7d4d88b4-a127-465d-b131-e37b28c46140/volumes" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.708411 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Jan 21 18:08:18 crc kubenswrapper[4808]: E0121 18:08:18.709409 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d4d88b4-a127-465d-b131-e37b28c46140" containerName="git-clone" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.709447 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d4d88b4-a127-465d-b131-e37b28c46140" containerName="git-clone" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.709668 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d4d88b4-a127-465d-b131-e37b28c46140" containerName="git-clone" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.711176 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.714487 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-4-sys-config" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.714520 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-4-ca" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.715102 4808 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-fbwgb" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.716548 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-4-global-ca" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.740911 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.775221 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8e5a3846-0833-4faf-a440-b1004dc941a2-build-system-configs\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.775284 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-container-storage-run\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.775330 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/8e5a3846-0833-4faf-a440-b1004dc941a2-builder-dockercfg-fbwgb-push\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.775359 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-build-blob-cache\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.775401 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/8e5a3846-0833-4faf-a440-b1004dc941a2-builder-dockercfg-fbwgb-pull\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.775422 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8e5a3846-0833-4faf-a440-b1004dc941a2-build-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.775508 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8e5a3846-0833-4faf-a440-b1004dc941a2-node-pullsecrets\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.775545 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8e5a3846-0833-4faf-a440-b1004dc941a2-build-proxy-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.775584 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmtzz\" (UniqueName: \"kubernetes.io/projected/8e5a3846-0833-4faf-a440-b1004dc941a2-kube-api-access-wmtzz\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.775608 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-buildworkdir\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.775671 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8e5a3846-0833-4faf-a440-b1004dc941a2-buildcachedir\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.775734 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-container-storage-root\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.877256 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8e5a3846-0833-4faf-a440-b1004dc941a2-build-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.877297 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/8e5a3846-0833-4faf-a440-b1004dc941a2-builder-dockercfg-fbwgb-pull\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.877333 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8e5a3846-0833-4faf-a440-b1004dc941a2-node-pullsecrets\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.877350 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8e5a3846-0833-4faf-a440-b1004dc941a2-build-proxy-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.877378 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmtzz\" (UniqueName: \"kubernetes.io/projected/8e5a3846-0833-4faf-a440-b1004dc941a2-kube-api-access-wmtzz\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.877398 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-buildworkdir\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.877422 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8e5a3846-0833-4faf-a440-b1004dc941a2-buildcachedir\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.877453 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-container-storage-root\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.877486 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8e5a3846-0833-4faf-a440-b1004dc941a2-build-system-configs\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.877510 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-container-storage-run\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.877529 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8e5a3846-0833-4faf-a440-b1004dc941a2-node-pullsecrets\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.877550 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/8e5a3846-0833-4faf-a440-b1004dc941a2-builder-dockercfg-fbwgb-push\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.877673 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-build-blob-cache\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.877726 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8e5a3846-0833-4faf-a440-b1004dc941a2-buildcachedir\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.878332 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-buildworkdir\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.878520 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8e5a3846-0833-4faf-a440-b1004dc941a2-build-system-configs\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.878516 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-container-storage-root\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.878572 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-build-blob-cache\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.878764 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-container-storage-run\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.878913 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8e5a3846-0833-4faf-a440-b1004dc941a2-build-proxy-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.879400 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8e5a3846-0833-4faf-a440-b1004dc941a2-build-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.882334 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/8e5a3846-0833-4faf-a440-b1004dc941a2-builder-dockercfg-fbwgb-push\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.882721 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/8e5a3846-0833-4faf-a440-b1004dc941a2-builder-dockercfg-fbwgb-pull\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:18 crc kubenswrapper[4808]: I0121 18:08:18.898523 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmtzz\" (UniqueName: \"kubernetes.io/projected/8e5a3846-0833-4faf-a440-b1004dc941a2-kube-api-access-wmtzz\") pod \"service-telemetry-operator-4-build\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:19 crc kubenswrapper[4808]: I0121 18:08:19.033086 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:19 crc kubenswrapper[4808]: I0121 18:08:19.596868 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Jan 21 18:08:20 crc kubenswrapper[4808]: I0121 18:08:20.121881 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-4-build" event={"ID":"8e5a3846-0833-4faf-a440-b1004dc941a2","Type":"ContainerStarted","Data":"30ea4ff8fb498da64555285e984f59c32f6636bb6511e150446450261061dd75"} Jan 21 18:08:23 crc kubenswrapper[4808]: I0121 18:08:23.153615 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-4-build" event={"ID":"8e5a3846-0833-4faf-a440-b1004dc941a2","Type":"ContainerStarted","Data":"b90527da25f4ab5f2a63e3ee252931a3c1f1fcdb25f601a9f97cb60a236b8bca"} Jan 21 18:08:23 crc kubenswrapper[4808]: E0121 18:08:23.216055 4808 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=7674227450993299356, SKID=, AKID=DF:1F:FD:7A:27:6A:84:51:E7:A3:2B:77:06:EA:03:AB:6E:1F:D3:D1 failed: x509: certificate signed by unknown authority" Jan 21 18:08:24 crc kubenswrapper[4808]: I0121 18:08:24.241504 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.169356 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-4-build" podUID="8e5a3846-0833-4faf-a440-b1004dc941a2" containerName="git-clone" containerID="cri-o://b90527da25f4ab5f2a63e3ee252931a3c1f1fcdb25f601a9f97cb60a236b8bca" gracePeriod=30 Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.585106 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-4-build_8e5a3846-0833-4faf-a440-b1004dc941a2/git-clone/0.log" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.585544 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.684503 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8e5a3846-0833-4faf-a440-b1004dc941a2-build-system-configs\") pod \"8e5a3846-0833-4faf-a440-b1004dc941a2\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.684582 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-container-storage-root\") pod \"8e5a3846-0833-4faf-a440-b1004dc941a2\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.684612 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8e5a3846-0833-4faf-a440-b1004dc941a2-node-pullsecrets\") pod \"8e5a3846-0833-4faf-a440-b1004dc941a2\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.684653 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/8e5a3846-0833-4faf-a440-b1004dc941a2-builder-dockercfg-fbwgb-pull\") pod \"8e5a3846-0833-4faf-a440-b1004dc941a2\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.684934 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmtzz\" (UniqueName: \"kubernetes.io/projected/8e5a3846-0833-4faf-a440-b1004dc941a2-kube-api-access-wmtzz\") pod \"8e5a3846-0833-4faf-a440-b1004dc941a2\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.685106 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/8e5a3846-0833-4faf-a440-b1004dc941a2-builder-dockercfg-fbwgb-push\") pod \"8e5a3846-0833-4faf-a440-b1004dc941a2\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.684981 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8e5a3846-0833-4faf-a440-b1004dc941a2-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "8e5a3846-0833-4faf-a440-b1004dc941a2" (UID: "8e5a3846-0833-4faf-a440-b1004dc941a2"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.685172 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-buildworkdir\") pod \"8e5a3846-0833-4faf-a440-b1004dc941a2\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.685234 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "8e5a3846-0833-4faf-a440-b1004dc941a2" (UID: "8e5a3846-0833-4faf-a440-b1004dc941a2"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.685329 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-build-blob-cache\") pod \"8e5a3846-0833-4faf-a440-b1004dc941a2\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.685385 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8e5a3846-0833-4faf-a440-b1004dc941a2-build-proxy-ca-bundles\") pod \"8e5a3846-0833-4faf-a440-b1004dc941a2\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.685436 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-container-storage-run\") pod \"8e5a3846-0833-4faf-a440-b1004dc941a2\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.685481 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8e5a3846-0833-4faf-a440-b1004dc941a2-buildcachedir\") pod \"8e5a3846-0833-4faf-a440-b1004dc941a2\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.685497 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e5a3846-0833-4faf-a440-b1004dc941a2-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "8e5a3846-0833-4faf-a440-b1004dc941a2" (UID: "8e5a3846-0833-4faf-a440-b1004dc941a2"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.685531 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8e5a3846-0833-4faf-a440-b1004dc941a2-build-ca-bundles\") pod \"8e5a3846-0833-4faf-a440-b1004dc941a2\" (UID: \"8e5a3846-0833-4faf-a440-b1004dc941a2\") " Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.685804 4808 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8e5a3846-0833-4faf-a440-b1004dc941a2-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.685806 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "8e5a3846-0833-4faf-a440-b1004dc941a2" (UID: "8e5a3846-0833-4faf-a440-b1004dc941a2"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.685817 4808 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.685871 4808 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8e5a3846-0833-4faf-a440-b1004dc941a2-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.686031 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "8e5a3846-0833-4faf-a440-b1004dc941a2" (UID: "8e5a3846-0833-4faf-a440-b1004dc941a2"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.686055 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8e5a3846-0833-4faf-a440-b1004dc941a2-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "8e5a3846-0833-4faf-a440-b1004dc941a2" (UID: "8e5a3846-0833-4faf-a440-b1004dc941a2"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.686581 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e5a3846-0833-4faf-a440-b1004dc941a2-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "8e5a3846-0833-4faf-a440-b1004dc941a2" (UID: "8e5a3846-0833-4faf-a440-b1004dc941a2"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.686622 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "8e5a3846-0833-4faf-a440-b1004dc941a2" (UID: "8e5a3846-0833-4faf-a440-b1004dc941a2"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.686816 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e5a3846-0833-4faf-a440-b1004dc941a2-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "8e5a3846-0833-4faf-a440-b1004dc941a2" (UID: "8e5a3846-0833-4faf-a440-b1004dc941a2"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.690069 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e5a3846-0833-4faf-a440-b1004dc941a2-builder-dockercfg-fbwgb-pull" (OuterVolumeSpecName: "builder-dockercfg-fbwgb-pull") pod "8e5a3846-0833-4faf-a440-b1004dc941a2" (UID: "8e5a3846-0833-4faf-a440-b1004dc941a2"). InnerVolumeSpecName "builder-dockercfg-fbwgb-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.690250 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e5a3846-0833-4faf-a440-b1004dc941a2-kube-api-access-wmtzz" (OuterVolumeSpecName: "kube-api-access-wmtzz") pod "8e5a3846-0833-4faf-a440-b1004dc941a2" (UID: "8e5a3846-0833-4faf-a440-b1004dc941a2"). InnerVolumeSpecName "kube-api-access-wmtzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.690781 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e5a3846-0833-4faf-a440-b1004dc941a2-builder-dockercfg-fbwgb-push" (OuterVolumeSpecName: "builder-dockercfg-fbwgb-push") pod "8e5a3846-0833-4faf-a440-b1004dc941a2" (UID: "8e5a3846-0833-4faf-a440-b1004dc941a2"). InnerVolumeSpecName "builder-dockercfg-fbwgb-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.786837 4808 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8e5a3846-0833-4faf-a440-b1004dc941a2-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.786885 4808 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8e5a3846-0833-4faf-a440-b1004dc941a2-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.786899 4808 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/8e5a3846-0833-4faf-a440-b1004dc941a2-builder-dockercfg-fbwgb-pull\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.787108 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmtzz\" (UniqueName: \"kubernetes.io/projected/8e5a3846-0833-4faf-a440-b1004dc941a2-kube-api-access-wmtzz\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.787121 4808 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/8e5a3846-0833-4faf-a440-b1004dc941a2-builder-dockercfg-fbwgb-push\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.787133 4808 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.787145 4808 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.787156 4808 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8e5a3846-0833-4faf-a440-b1004dc941a2-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:25 crc kubenswrapper[4808]: I0121 18:08:25.787167 4808 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8e5a3846-0833-4faf-a440-b1004dc941a2-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:26 crc kubenswrapper[4808]: I0121 18:08:26.178096 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-4-build_8e5a3846-0833-4faf-a440-b1004dc941a2/git-clone/0.log" Jan 21 18:08:26 crc kubenswrapper[4808]: I0121 18:08:26.178157 4808 generic.go:334] "Generic (PLEG): container finished" podID="8e5a3846-0833-4faf-a440-b1004dc941a2" containerID="b90527da25f4ab5f2a63e3ee252931a3c1f1fcdb25f601a9f97cb60a236b8bca" exitCode=1 Jan 21 18:08:26 crc kubenswrapper[4808]: I0121 18:08:26.178186 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-4-build" event={"ID":"8e5a3846-0833-4faf-a440-b1004dc941a2","Type":"ContainerDied","Data":"b90527da25f4ab5f2a63e3ee252931a3c1f1fcdb25f601a9f97cb60a236b8bca"} Jan 21 18:08:26 crc kubenswrapper[4808]: I0121 18:08:26.178218 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-4-build" event={"ID":"8e5a3846-0833-4faf-a440-b1004dc941a2","Type":"ContainerDied","Data":"30ea4ff8fb498da64555285e984f59c32f6636bb6511e150446450261061dd75"} Jan 21 18:08:26 crc kubenswrapper[4808]: I0121 18:08:26.178262 4808 scope.go:117] "RemoveContainer" containerID="b90527da25f4ab5f2a63e3ee252931a3c1f1fcdb25f601a9f97cb60a236b8bca" Jan 21 18:08:26 crc kubenswrapper[4808]: I0121 18:08:26.178289 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-4-build" Jan 21 18:08:26 crc kubenswrapper[4808]: I0121 18:08:26.201754 4808 scope.go:117] "RemoveContainer" containerID="b90527da25f4ab5f2a63e3ee252931a3c1f1fcdb25f601a9f97cb60a236b8bca" Jan 21 18:08:26 crc kubenswrapper[4808]: E0121 18:08:26.202396 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b90527da25f4ab5f2a63e3ee252931a3c1f1fcdb25f601a9f97cb60a236b8bca\": container with ID starting with b90527da25f4ab5f2a63e3ee252931a3c1f1fcdb25f601a9f97cb60a236b8bca not found: ID does not exist" containerID="b90527da25f4ab5f2a63e3ee252931a3c1f1fcdb25f601a9f97cb60a236b8bca" Jan 21 18:08:26 crc kubenswrapper[4808]: I0121 18:08:26.202448 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b90527da25f4ab5f2a63e3ee252931a3c1f1fcdb25f601a9f97cb60a236b8bca"} err="failed to get container status \"b90527da25f4ab5f2a63e3ee252931a3c1f1fcdb25f601a9f97cb60a236b8bca\": rpc error: code = NotFound desc = could not find container \"b90527da25f4ab5f2a63e3ee252931a3c1f1fcdb25f601a9f97cb60a236b8bca\": container with ID starting with b90527da25f4ab5f2a63e3ee252931a3c1f1fcdb25f601a9f97cb60a236b8bca not found: ID does not exist" Jan 21 18:08:26 crc kubenswrapper[4808]: I0121 18:08:26.235153 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Jan 21 18:08:26 crc kubenswrapper[4808]: I0121 18:08:26.245026 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Jan 21 18:08:27 crc kubenswrapper[4808]: I0121 18:08:27.545165 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e5a3846-0833-4faf-a440-b1004dc941a2" path="/var/lib/kubelet/pods/8e5a3846-0833-4faf-a440-b1004dc941a2/volumes" Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.780622 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Jan 21 18:08:35 crc kubenswrapper[4808]: E0121 18:08:35.781531 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e5a3846-0833-4faf-a440-b1004dc941a2" containerName="git-clone" Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.781551 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e5a3846-0833-4faf-a440-b1004dc941a2" containerName="git-clone" Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.781724 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e5a3846-0833-4faf-a440-b1004dc941a2" containerName="git-clone" Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.783047 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.784473 4808 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-fbwgb" Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.786272 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-5-ca" Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.787604 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-5-global-ca" Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.789233 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-5-sys-config" Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.802470 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.935863 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgvd2\" (UniqueName: \"kubernetes.io/projected/9d48a05d-e35f-4955-83db-ee2f6aaf851a-kube-api-access-mgvd2\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.936254 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-blob-cache\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.936374 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9d48a05d-e35f-4955-83db-ee2f6aaf851a-node-pullsecrets\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.936479 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9d48a05d-e35f-4955-83db-ee2f6aaf851a-buildcachedir\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.936596 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-container-storage-run\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.936716 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-proxy-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.936821 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-system-configs\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.936905 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/9d48a05d-e35f-4955-83db-ee2f6aaf851a-builder-dockercfg-fbwgb-pull\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.936976 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.937042 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-buildworkdir\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.937077 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/9d48a05d-e35f-4955-83db-ee2f6aaf851a-builder-dockercfg-fbwgb-push\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:35 crc kubenswrapper[4808]: I0121 18:08:35.937116 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-container-storage-root\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.038011 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-container-storage-run\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.038102 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-proxy-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.038128 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-system-configs\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.038161 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/9d48a05d-e35f-4955-83db-ee2f6aaf851a-builder-dockercfg-fbwgb-pull\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.038193 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.038254 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-buildworkdir\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.038301 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/9d48a05d-e35f-4955-83db-ee2f6aaf851a-builder-dockercfg-fbwgb-push\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.038321 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-container-storage-root\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.038361 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgvd2\" (UniqueName: \"kubernetes.io/projected/9d48a05d-e35f-4955-83db-ee2f6aaf851a-kube-api-access-mgvd2\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.038383 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-blob-cache\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.038402 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9d48a05d-e35f-4955-83db-ee2f6aaf851a-node-pullsecrets\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.038424 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9d48a05d-e35f-4955-83db-ee2f6aaf851a-buildcachedir\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.038515 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9d48a05d-e35f-4955-83db-ee2f6aaf851a-buildcachedir\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.038809 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9d48a05d-e35f-4955-83db-ee2f6aaf851a-node-pullsecrets\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.039135 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-blob-cache\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.039263 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-container-storage-root\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.039677 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-buildworkdir\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.039796 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.040274 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-proxy-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.040753 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-container-storage-run\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.041347 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-system-configs\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.046161 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/9d48a05d-e35f-4955-83db-ee2f6aaf851a-builder-dockercfg-fbwgb-pull\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.046175 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/9d48a05d-e35f-4955-83db-ee2f6aaf851a-builder-dockercfg-fbwgb-push\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.053476 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgvd2\" (UniqueName: \"kubernetes.io/projected/9d48a05d-e35f-4955-83db-ee2f6aaf851a-kube-api-access-mgvd2\") pod \"service-telemetry-operator-5-build\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.105147 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:36 crc kubenswrapper[4808]: I0121 18:08:36.329812 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Jan 21 18:08:37 crc kubenswrapper[4808]: I0121 18:08:37.260886 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-5-build" event={"ID":"9d48a05d-e35f-4955-83db-ee2f6aaf851a","Type":"ContainerStarted","Data":"5ff67257177c1295055eee0368f9dc82f299c0cd7395c4e3cd2e549956f499b3"} Jan 21 18:08:37 crc kubenswrapper[4808]: I0121 18:08:37.261202 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-5-build" event={"ID":"9d48a05d-e35f-4955-83db-ee2f6aaf851a","Type":"ContainerStarted","Data":"7b39f82f276fd0dfaab7c4c37a410bacda64764e8778b36066a18b97d5aed179"} Jan 21 18:08:37 crc kubenswrapper[4808]: E0121 18:08:37.319055 4808 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=7674227450993299356, SKID=, AKID=DF:1F:FD:7A:27:6A:84:51:E7:A3:2B:77:06:EA:03:AB:6E:1F:D3:D1 failed: x509: certificate signed by unknown authority" Jan 21 18:08:38 crc kubenswrapper[4808]: I0121 18:08:38.346298 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.277080 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-5-build" podUID="9d48a05d-e35f-4955-83db-ee2f6aaf851a" containerName="git-clone" containerID="cri-o://5ff67257177c1295055eee0368f9dc82f299c0cd7395c4e3cd2e549956f499b3" gracePeriod=30 Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.673027 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-5-build_9d48a05d-e35f-4955-83db-ee2f6aaf851a/git-clone/0.log" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.673284 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.798120 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/9d48a05d-e35f-4955-83db-ee2f6aaf851a-builder-dockercfg-fbwgb-pull\") pod \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.798469 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/9d48a05d-e35f-4955-83db-ee2f6aaf851a-builder-dockercfg-fbwgb-push\") pod \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.798522 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-proxy-ca-bundles\") pod \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.798564 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-blob-cache\") pod \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.798590 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-container-storage-run\") pod \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.798615 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-container-storage-root\") pod \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.798648 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-buildworkdir\") pod \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.799021 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "9d48a05d-e35f-4955-83db-ee2f6aaf851a" (UID: "9d48a05d-e35f-4955-83db-ee2f6aaf851a"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.799057 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "9d48a05d-e35f-4955-83db-ee2f6aaf851a" (UID: "9d48a05d-e35f-4955-83db-ee2f6aaf851a"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.799074 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "9d48a05d-e35f-4955-83db-ee2f6aaf851a" (UID: "9d48a05d-e35f-4955-83db-ee2f6aaf851a"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.799263 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "9d48a05d-e35f-4955-83db-ee2f6aaf851a" (UID: "9d48a05d-e35f-4955-83db-ee2f6aaf851a"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.799150 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-ca-bundles\") pod \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.799453 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "9d48a05d-e35f-4955-83db-ee2f6aaf851a" (UID: "9d48a05d-e35f-4955-83db-ee2f6aaf851a"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.799716 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "9d48a05d-e35f-4955-83db-ee2f6aaf851a" (UID: "9d48a05d-e35f-4955-83db-ee2f6aaf851a"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.799889 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgvd2\" (UniqueName: \"kubernetes.io/projected/9d48a05d-e35f-4955-83db-ee2f6aaf851a-kube-api-access-mgvd2\") pod \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.799931 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-system-configs\") pod \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.800000 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9d48a05d-e35f-4955-83db-ee2f6aaf851a-node-pullsecrets\") pod \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.800076 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9d48a05d-e35f-4955-83db-ee2f6aaf851a-buildcachedir\") pod \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\" (UID: \"9d48a05d-e35f-4955-83db-ee2f6aaf851a\") " Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.800148 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9d48a05d-e35f-4955-83db-ee2f6aaf851a-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "9d48a05d-e35f-4955-83db-ee2f6aaf851a" (UID: "9d48a05d-e35f-4955-83db-ee2f6aaf851a"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.800264 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9d48a05d-e35f-4955-83db-ee2f6aaf851a-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "9d48a05d-e35f-4955-83db-ee2f6aaf851a" (UID: "9d48a05d-e35f-4955-83db-ee2f6aaf851a"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.800582 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "9d48a05d-e35f-4955-83db-ee2f6aaf851a" (UID: "9d48a05d-e35f-4955-83db-ee2f6aaf851a"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.800933 4808 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.800970 4808 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.800990 4808 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.801006 4808 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.801022 4808 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9d48a05d-e35f-4955-83db-ee2f6aaf851a-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.801036 4808 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.801052 4808 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9d48a05d-e35f-4955-83db-ee2f6aaf851a-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.801066 4808 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9d48a05d-e35f-4955-83db-ee2f6aaf851a-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.801080 4808 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9d48a05d-e35f-4955-83db-ee2f6aaf851a-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.803082 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d48a05d-e35f-4955-83db-ee2f6aaf851a-builder-dockercfg-fbwgb-pull" (OuterVolumeSpecName: "builder-dockercfg-fbwgb-pull") pod "9d48a05d-e35f-4955-83db-ee2f6aaf851a" (UID: "9d48a05d-e35f-4955-83db-ee2f6aaf851a"). InnerVolumeSpecName "builder-dockercfg-fbwgb-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.803321 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d48a05d-e35f-4955-83db-ee2f6aaf851a-builder-dockercfg-fbwgb-push" (OuterVolumeSpecName: "builder-dockercfg-fbwgb-push") pod "9d48a05d-e35f-4955-83db-ee2f6aaf851a" (UID: "9d48a05d-e35f-4955-83db-ee2f6aaf851a"). InnerVolumeSpecName "builder-dockercfg-fbwgb-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.803370 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d48a05d-e35f-4955-83db-ee2f6aaf851a-kube-api-access-mgvd2" (OuterVolumeSpecName: "kube-api-access-mgvd2") pod "9d48a05d-e35f-4955-83db-ee2f6aaf851a" (UID: "9d48a05d-e35f-4955-83db-ee2f6aaf851a"). InnerVolumeSpecName "kube-api-access-mgvd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.901978 4808 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-fbwgb-pull\" (UniqueName: \"kubernetes.io/secret/9d48a05d-e35f-4955-83db-ee2f6aaf851a-builder-dockercfg-fbwgb-pull\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.902009 4808 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-fbwgb-push\" (UniqueName: \"kubernetes.io/secret/9d48a05d-e35f-4955-83db-ee2f6aaf851a-builder-dockercfg-fbwgb-push\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:39 crc kubenswrapper[4808]: I0121 18:08:39.902020 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgvd2\" (UniqueName: \"kubernetes.io/projected/9d48a05d-e35f-4955-83db-ee2f6aaf851a-kube-api-access-mgvd2\") on node \"crc\" DevicePath \"\"" Jan 21 18:08:40 crc kubenswrapper[4808]: I0121 18:08:40.287773 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-5-build_9d48a05d-e35f-4955-83db-ee2f6aaf851a/git-clone/0.log" Jan 21 18:08:40 crc kubenswrapper[4808]: I0121 18:08:40.287817 4808 generic.go:334] "Generic (PLEG): container finished" podID="9d48a05d-e35f-4955-83db-ee2f6aaf851a" containerID="5ff67257177c1295055eee0368f9dc82f299c0cd7395c4e3cd2e549956f499b3" exitCode=1 Jan 21 18:08:40 crc kubenswrapper[4808]: I0121 18:08:40.287844 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-5-build" event={"ID":"9d48a05d-e35f-4955-83db-ee2f6aaf851a","Type":"ContainerDied","Data":"5ff67257177c1295055eee0368f9dc82f299c0cd7395c4e3cd2e549956f499b3"} Jan 21 18:08:40 crc kubenswrapper[4808]: I0121 18:08:40.287870 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-5-build" event={"ID":"9d48a05d-e35f-4955-83db-ee2f6aaf851a","Type":"ContainerDied","Data":"7b39f82f276fd0dfaab7c4c37a410bacda64764e8778b36066a18b97d5aed179"} Jan 21 18:08:40 crc kubenswrapper[4808]: I0121 18:08:40.287888 4808 scope.go:117] "RemoveContainer" containerID="5ff67257177c1295055eee0368f9dc82f299c0cd7395c4e3cd2e549956f499b3" Jan 21 18:08:40 crc kubenswrapper[4808]: I0121 18:08:40.287922 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-5-build" Jan 21 18:08:40 crc kubenswrapper[4808]: I0121 18:08:40.325978 4808 scope.go:117] "RemoveContainer" containerID="5ff67257177c1295055eee0368f9dc82f299c0cd7395c4e3cd2e549956f499b3" Jan 21 18:08:40 crc kubenswrapper[4808]: E0121 18:08:40.326518 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ff67257177c1295055eee0368f9dc82f299c0cd7395c4e3cd2e549956f499b3\": container with ID starting with 5ff67257177c1295055eee0368f9dc82f299c0cd7395c4e3cd2e549956f499b3 not found: ID does not exist" containerID="5ff67257177c1295055eee0368f9dc82f299c0cd7395c4e3cd2e549956f499b3" Jan 21 18:08:40 crc kubenswrapper[4808]: I0121 18:08:40.326572 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ff67257177c1295055eee0368f9dc82f299c0cd7395c4e3cd2e549956f499b3"} err="failed to get container status \"5ff67257177c1295055eee0368f9dc82f299c0cd7395c4e3cd2e549956f499b3\": rpc error: code = NotFound desc = could not find container \"5ff67257177c1295055eee0368f9dc82f299c0cd7395c4e3cd2e549956f499b3\": container with ID starting with 5ff67257177c1295055eee0368f9dc82f299c0cd7395c4e3cd2e549956f499b3 not found: ID does not exist" Jan 21 18:08:40 crc kubenswrapper[4808]: I0121 18:08:40.358750 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Jan 21 18:08:40 crc kubenswrapper[4808]: I0121 18:08:40.366040 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Jan 21 18:08:41 crc kubenswrapper[4808]: I0121 18:08:41.531225 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d48a05d-e35f-4955-83db-ee2f6aaf851a" path="/var/lib/kubelet/pods/9d48a05d-e35f-4955-83db-ee2f6aaf851a/volumes" Jan 21 18:08:45 crc kubenswrapper[4808]: I0121 18:08:45.931436 4808 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 21 18:08:58 crc kubenswrapper[4808]: I0121 18:08:58.332032 4808 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-operators-wbqf6" podUID="37705e9c-abf0-42ac-8fc3-03e0b2fa732a" containerName="registry-server" probeResult="failure" output=< Jan 21 18:08:58 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Jan 21 18:08:58 crc kubenswrapper[4808]: > Jan 21 18:08:58 crc kubenswrapper[4808]: I0121 18:08:58.398839 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-operators-wbqf6" podUID="37705e9c-abf0-42ac-8fc3-03e0b2fa732a" containerName="registry-server" probeResult="failure" output=< Jan 21 18:08:58 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Jan 21 18:08:58 crc kubenswrapper[4808]: > Jan 21 18:08:59 crc kubenswrapper[4808]: I0121 18:08:59.599354 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:08:59 crc kubenswrapper[4808]: I0121 18:08:59.600306 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:09:17 crc kubenswrapper[4808]: I0121 18:09:17.660937 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qk2zk/must-gather-kr5nm"] Jan 21 18:09:17 crc kubenswrapper[4808]: E0121 18:09:17.663733 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d48a05d-e35f-4955-83db-ee2f6aaf851a" containerName="git-clone" Jan 21 18:09:17 crc kubenswrapper[4808]: I0121 18:09:17.663847 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d48a05d-e35f-4955-83db-ee2f6aaf851a" containerName="git-clone" Jan 21 18:09:17 crc kubenswrapper[4808]: I0121 18:09:17.664070 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d48a05d-e35f-4955-83db-ee2f6aaf851a" containerName="git-clone" Jan 21 18:09:17 crc kubenswrapper[4808]: I0121 18:09:17.665378 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qk2zk/must-gather-kr5nm" Jan 21 18:09:17 crc kubenswrapper[4808]: I0121 18:09:17.668550 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qk2zk"/"kube-root-ca.crt" Jan 21 18:09:17 crc kubenswrapper[4808]: I0121 18:09:17.669053 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qk2zk"/"openshift-service-ca.crt" Jan 21 18:09:17 crc kubenswrapper[4808]: I0121 18:09:17.674584 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-qk2zk"/"default-dockercfg-hxd5c" Jan 21 18:09:17 crc kubenswrapper[4808]: I0121 18:09:17.681608 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qk2zk/must-gather-kr5nm"] Jan 21 18:09:17 crc kubenswrapper[4808]: I0121 18:09:17.777037 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c892c4a1-e11b-4439-af21-77fcaa805676-must-gather-output\") pod \"must-gather-kr5nm\" (UID: \"c892c4a1-e11b-4439-af21-77fcaa805676\") " pod="openshift-must-gather-qk2zk/must-gather-kr5nm" Jan 21 18:09:17 crc kubenswrapper[4808]: I0121 18:09:17.777624 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpjpw\" (UniqueName: \"kubernetes.io/projected/c892c4a1-e11b-4439-af21-77fcaa805676-kube-api-access-hpjpw\") pod \"must-gather-kr5nm\" (UID: \"c892c4a1-e11b-4439-af21-77fcaa805676\") " pod="openshift-must-gather-qk2zk/must-gather-kr5nm" Jan 21 18:09:17 crc kubenswrapper[4808]: I0121 18:09:17.879017 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c892c4a1-e11b-4439-af21-77fcaa805676-must-gather-output\") pod \"must-gather-kr5nm\" (UID: \"c892c4a1-e11b-4439-af21-77fcaa805676\") " pod="openshift-must-gather-qk2zk/must-gather-kr5nm" Jan 21 18:09:17 crc kubenswrapper[4808]: I0121 18:09:17.879103 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpjpw\" (UniqueName: \"kubernetes.io/projected/c892c4a1-e11b-4439-af21-77fcaa805676-kube-api-access-hpjpw\") pod \"must-gather-kr5nm\" (UID: \"c892c4a1-e11b-4439-af21-77fcaa805676\") " pod="openshift-must-gather-qk2zk/must-gather-kr5nm" Jan 21 18:09:17 crc kubenswrapper[4808]: I0121 18:09:17.879589 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c892c4a1-e11b-4439-af21-77fcaa805676-must-gather-output\") pod \"must-gather-kr5nm\" (UID: \"c892c4a1-e11b-4439-af21-77fcaa805676\") " pod="openshift-must-gather-qk2zk/must-gather-kr5nm" Jan 21 18:09:17 crc kubenswrapper[4808]: I0121 18:09:17.899915 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpjpw\" (UniqueName: \"kubernetes.io/projected/c892c4a1-e11b-4439-af21-77fcaa805676-kube-api-access-hpjpw\") pod \"must-gather-kr5nm\" (UID: \"c892c4a1-e11b-4439-af21-77fcaa805676\") " pod="openshift-must-gather-qk2zk/must-gather-kr5nm" Jan 21 18:09:17 crc kubenswrapper[4808]: I0121 18:09:17.990124 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qk2zk/must-gather-kr5nm" Jan 21 18:09:18 crc kubenswrapper[4808]: I0121 18:09:18.483759 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qk2zk/must-gather-kr5nm"] Jan 21 18:09:18 crc kubenswrapper[4808]: I0121 18:09:18.568031 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qk2zk/must-gather-kr5nm" event={"ID":"c892c4a1-e11b-4439-af21-77fcaa805676","Type":"ContainerStarted","Data":"c624933940bf65c4ff4c46067c809581f40c82abf0ac75266535273db7027d90"} Jan 21 18:09:26 crc kubenswrapper[4808]: I0121 18:09:26.648923 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qk2zk/must-gather-kr5nm" event={"ID":"c892c4a1-e11b-4439-af21-77fcaa805676","Type":"ContainerStarted","Data":"aa293fe7a8d6e70fe968edf4c927bc4fe47cefd01240451abe7449b0914641af"} Jan 21 18:09:26 crc kubenswrapper[4808]: I0121 18:09:26.649614 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qk2zk/must-gather-kr5nm" event={"ID":"c892c4a1-e11b-4439-af21-77fcaa805676","Type":"ContainerStarted","Data":"0faf1d78f309b620b407cecbd781576d1a7d417a36e04a3936cee00c14c61503"} Jan 21 18:09:26 crc kubenswrapper[4808]: I0121 18:09:26.669362 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-qk2zk/must-gather-kr5nm" podStartSLOduration=2.732985553 podStartE2EDuration="9.669345984s" podCreationTimestamp="2026-01-21 18:09:17 +0000 UTC" firstStartedPulling="2026-01-21 18:09:18.491690772 +0000 UTC m=+793.222201657" lastFinishedPulling="2026-01-21 18:09:25.428051203 +0000 UTC m=+800.158562088" observedRunningTime="2026-01-21 18:09:26.668017473 +0000 UTC m=+801.398528368" watchObservedRunningTime="2026-01-21 18:09:26.669345984 +0000 UTC m=+801.399856869" Jan 21 18:09:29 crc kubenswrapper[4808]: I0121 18:09:29.600285 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:09:29 crc kubenswrapper[4808]: I0121 18:09:29.600687 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:09:37 crc kubenswrapper[4808]: I0121 18:09:37.849898 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-jl78z_11782075-8384-415d-a1b9-6c972f8f6f66/control-plane-machine-set-operator/0.log" Jan 21 18:09:37 crc kubenswrapper[4808]: I0121 18:09:37.866448 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4g7bd_4c87e4f3-6255-4fe3-a52c-726a1acf7ba8/kube-rbac-proxy/0.log" Jan 21 18:09:37 crc kubenswrapper[4808]: I0121 18:09:37.872919 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4g7bd_4c87e4f3-6255-4fe3-a52c-726a1acf7ba8/machine-api-operator/0.log" Jan 21 18:09:42 crc kubenswrapper[4808]: I0121 18:09:42.454489 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-mvjvw_045426c7-6605-4623-93c3-1575773a5ecc/cert-manager-controller/0.log" Jan 21 18:09:42 crc kubenswrapper[4808]: I0121 18:09:42.468410 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-8jch4_71092cb7-08bc-4f0f-9696-838e2fc52d62/cert-manager-cainjector/0.log" Jan 21 18:09:42 crc kubenswrapper[4808]: I0121 18:09:42.480563 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-db8jl_d30e2538-925e-4c64-8d47-7d0b7255f69d/cert-manager-webhook/0.log" Jan 21 18:09:47 crc kubenswrapper[4808]: I0121 18:09:47.208983 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-6mhk8_14e13b58-9524-4ad9-94db-848ef5698394/prometheus-operator/0.log" Jan 21 18:09:47 crc kubenswrapper[4808]: I0121 18:09:47.220193 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr_5e732623-a8de-4909-850b-6728c2db0689/prometheus-operator-admission-webhook/0.log" Jan 21 18:09:47 crc kubenswrapper[4808]: I0121 18:09:47.236555 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-59957c94cc-dss74_a0ebf276-ed9c-4cb7-9423-0089c9e08533/prometheus-operator-admission-webhook/0.log" Jan 21 18:09:47 crc kubenswrapper[4808]: I0121 18:09:47.251827 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-fh9px_a54735e0-9845-41e8-a765-cb6a4d04943c/operator/0.log" Jan 21 18:09:47 crc kubenswrapper[4808]: I0121 18:09:47.264094 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-lkgq6_f24a8af6-f286-470e-a41e-86d9cdd38963/perses-operator/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.152451 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d_a8baaf63-dc73-4834-870d-bb3c7cd44ede/extract/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.162788 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d_a8baaf63-dc73-4834-870d-bb3c7cd44ede/util/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.189578 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931amlk5d_a8baaf63-dc73-4834-870d-bb3c7cd44ede/pull/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.200425 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm_8935b136-3537-4e51-ae90-9c3729258689/extract/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.207600 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm_8935b136-3537-4e51-ae90-9c3729258689/util/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.216155 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f8tcgm_8935b136-3537-4e51-ae90-9c3729258689/pull/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.228155 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9_61b09e55-6f9f-4dd0-a174-8fcd696335c2/extract/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.237217 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9_61b09e55-6f9f-4dd0-a174-8fcd696335c2/util/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.246834 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5emsgv9_61b09e55-6f9f-4dd0-a174-8fcd696335c2/pull/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.259122 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd_de6d62bf-8431-46c9-9720-9874644fabd9/extract/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.265225 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd_de6d62bf-8431-46c9-9720-9874644fabd9/util/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.273955 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08645sd_de6d62bf-8431-46c9-9720-9874644fabd9/pull/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.290522 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4pfvs_9d1c03b0-e10f-47ae-841c-153e4ae0f9cd/registry-server/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.295976 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4pfvs_9d1c03b0-e10f-47ae-841c-153e4ae0f9cd/extract-utilities/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.307524 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4pfvs_9d1c03b0-e10f-47ae-841c-153e4ae0f9cd/extract-content/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.331810 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5c9xs_21ebff69-0bfe-427d-8330-4e454159dd01/registry-server/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.342375 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5c9xs_21ebff69-0bfe-427d-8330-4e454159dd01/extract-utilities/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.351342 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5c9xs_21ebff69-0bfe-427d-8330-4e454159dd01/extract-content/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.370925 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6m94d_62cdb877-c15b-4990-8820-3dcc6ffc0b04/registry-server/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.377566 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6m94d_62cdb877-c15b-4990-8820-3dcc6ffc0b04/extract-utilities/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.385772 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6m94d_62cdb877-c15b-4990-8820-3dcc6ffc0b04/extract-content/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.403401 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cbwpb_1df03daa-d900-4a6d-9776-26f3b05843ce/registry-server/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.408054 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cbwpb_1df03daa-d900-4a6d-9776-26f3b05843ce/extract-utilities/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.418948 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cbwpb_1df03daa-d900-4a6d-9776-26f3b05843ce/extract-content/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.454531 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hzsq2_7f554b03-f857-4309-ba4c-91e8c607a14d/registry-server/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.458983 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hzsq2_7f554b03-f857-4309-ba4c-91e8c607a14d/extract-utilities/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.467194 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hzsq2_7f554b03-f857-4309-ba4c-91e8c607a14d/extract-content/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.488230 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kd6dv_dde2d0c2-70a4-4d0a-bc7a-29fd72c50095/registry-server/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.493177 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kd6dv_dde2d0c2-70a4-4d0a-bc7a-29fd72c50095/extract-utilities/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.500375 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kd6dv_dde2d0c2-70a4-4d0a-bc7a-29fd72c50095/extract-content/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.517603 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqzlp_c5322a4c-c457-48c5-9c8d-6ce242b4b87b/registry-server/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.524676 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqzlp_c5322a4c-c457-48c5-9c8d-6ce242b4b87b/extract-utilities/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.533042 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nqzlp_c5322a4c-c457-48c5-9c8d-6ce242b4b87b/extract-content/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.664982 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xstv7_173b6b54-1024-4806-8056-985010126525/registry-server/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.669760 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xstv7_173b6b54-1024-4806-8056-985010126525/extract-utilities/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.678528 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xstv7_173b6b54-1024-4806-8056-985010126525/extract-content/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.962386 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mzpzk_688d7727-51ad-456f-a651-cea655ee7507/registry-server/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.968879 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mzpzk_688d7727-51ad-456f-a651-cea655ee7507/extract-utilities/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.977497 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mzpzk_688d7727-51ad-456f-a651-cea655ee7507/extract-content/0.log" Jan 21 18:09:52 crc kubenswrapper[4808]: I0121 18:09:52.996929 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-cklmp_0a6a5330-18dc-4ab0-bfb0-6b98549e28bc/marketplace-operator/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.012032 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2rh88_541d9739-547c-4697-9658-b29bc90e20c6/registry-server/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.017023 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2rh88_541d9739-547c-4697-9658-b29bc90e20c6/extract-utilities/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.024278 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2rh88_541d9739-547c-4697-9658-b29bc90e20c6/extract-content/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.042727 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-54shl_9c4c6107-02cd-47e8-b29a-88190d15f030/registry-server/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.047493 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-54shl_9c4c6107-02cd-47e8-b29a-88190d15f030/extract-utilities/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.056051 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-54shl_9c4c6107-02cd-47e8-b29a-88190d15f030/extract-content/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.095323 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-59dvj_bc01e22f-4eb3-46d9-8bba-faffce8f294e/registry-server/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.100974 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-59dvj_bc01e22f-4eb3-46d9-8bba-faffce8f294e/extract-utilities/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.108498 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-59dvj_bc01e22f-4eb3-46d9-8bba-faffce8f294e/extract-content/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.129706 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bwpzc_e41b3557-3ab9-43d1-b895-53834748a458/registry-server/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.135510 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bwpzc_e41b3557-3ab9-43d1-b895-53834748a458/extract-utilities/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.148217 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bwpzc_e41b3557-3ab9-43d1-b895-53834748a458/extract-content/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.165568 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hvw6j_e892666d-d0e9-4169-98d7-70687ac29c1b/registry-server/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.171606 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hvw6j_e892666d-d0e9-4169-98d7-70687ac29c1b/extract-utilities/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.179421 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hvw6j_e892666d-d0e9-4169-98d7-70687ac29c1b/extract-content/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.196804 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jl45c_ff29d64c-c8c9-4cb7-9a58-4f04f78034e2/registry-server/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.203281 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jl45c_ff29d64c-c8c9-4cb7-9a58-4f04f78034e2/extract-utilities/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.210850 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jl45c_ff29d64c-c8c9-4cb7-9a58-4f04f78034e2/extract-content/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.229216 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tj8vd_afc0321c-4482-40a4-92e3-67f5f78f5c6e/registry-server/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.239345 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tj8vd_afc0321c-4482-40a4-92e3-67f5f78f5c6e/extract-utilities/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.246747 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tj8vd_afc0321c-4482-40a4-92e3-67f5f78f5c6e/extract-content/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.312661 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wbqf6_37705e9c-abf0-42ac-8fc3-03e0b2fa732a/registry-server/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.317648 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wbqf6_37705e9c-abf0-42ac-8fc3-03e0b2fa732a/extract-utilities/0.log" Jan 21 18:09:53 crc kubenswrapper[4808]: I0121 18:09:53.325296 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wbqf6_37705e9c-abf0-42ac-8fc3-03e0b2fa732a/extract-content/0.log" Jan 21 18:09:57 crc kubenswrapper[4808]: I0121 18:09:57.009309 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-6mhk8_14e13b58-9524-4ad9-94db-848ef5698394/prometheus-operator/0.log" Jan 21 18:09:57 crc kubenswrapper[4808]: I0121 18:09:57.024546 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr_5e732623-a8de-4909-850b-6728c2db0689/prometheus-operator-admission-webhook/0.log" Jan 21 18:09:57 crc kubenswrapper[4808]: I0121 18:09:57.037885 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-59957c94cc-dss74_a0ebf276-ed9c-4cb7-9423-0089c9e08533/prometheus-operator-admission-webhook/0.log" Jan 21 18:09:57 crc kubenswrapper[4808]: I0121 18:09:57.061385 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-fh9px_a54735e0-9845-41e8-a765-cb6a4d04943c/operator/0.log" Jan 21 18:09:57 crc kubenswrapper[4808]: I0121 18:09:57.074611 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-lkgq6_f24a8af6-f286-470e-a41e-86d9cdd38963/perses-operator/0.log" Jan 21 18:09:59 crc kubenswrapper[4808]: I0121 18:09:59.600369 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:09:59 crc kubenswrapper[4808]: I0121 18:09:59.600451 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:09:59 crc kubenswrapper[4808]: I0121 18:09:59.600513 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 18:09:59 crc kubenswrapper[4808]: I0121 18:09:59.601334 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5fd5f938b5c81db84b9cd3776775f788a436e3c24fc43c86adbeb98927f5200d"} pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 18:09:59 crc kubenswrapper[4808]: I0121 18:09:59.601432 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" containerID="cri-o://5fd5f938b5c81db84b9cd3776775f788a436e3c24fc43c86adbeb98927f5200d" gracePeriod=600 Jan 21 18:10:01 crc kubenswrapper[4808]: I0121 18:10:01.878149 4808 generic.go:334] "Generic (PLEG): container finished" podID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerID="5fd5f938b5c81db84b9cd3776775f788a436e3c24fc43c86adbeb98927f5200d" exitCode=0 Jan 21 18:10:01 crc kubenswrapper[4808]: I0121 18:10:01.878284 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerDied","Data":"5fd5f938b5c81db84b9cd3776775f788a436e3c24fc43c86adbeb98927f5200d"} Jan 21 18:10:01 crc kubenswrapper[4808]: I0121 18:10:01.878576 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerStarted","Data":"22ca6704d095a7830b6b1f1aa076b2fe0d7fab0a47c466111f510e13e0730cbe"} Jan 21 18:10:01 crc kubenswrapper[4808]: I0121 18:10:01.878607 4808 scope.go:117] "RemoveContainer" containerID="00f0bd827addaebc04ffc5544dc603b3fe979bfa986819c93eca5d866ec75282" Jan 21 18:10:05 crc kubenswrapper[4808]: I0121 18:10:05.015099 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-6mhk8_14e13b58-9524-4ad9-94db-848ef5698394/prometheus-operator/0.log" Jan 21 18:10:05 crc kubenswrapper[4808]: I0121 18:10:05.037981 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-59957c94cc-5w9qr_5e732623-a8de-4909-850b-6728c2db0689/prometheus-operator-admission-webhook/0.log" Jan 21 18:10:05 crc kubenswrapper[4808]: I0121 18:10:05.053166 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-59957c94cc-dss74_a0ebf276-ed9c-4cb7-9423-0089c9e08533/prometheus-operator-admission-webhook/0.log" Jan 21 18:10:05 crc kubenswrapper[4808]: I0121 18:10:05.071258 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-fh9px_a54735e0-9845-41e8-a765-cb6a4d04943c/operator/0.log" Jan 21 18:10:05 crc kubenswrapper[4808]: I0121 18:10:05.087796 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-lkgq6_f24a8af6-f286-470e-a41e-86d9cdd38963/perses-operator/0.log" Jan 21 18:10:05 crc kubenswrapper[4808]: I0121 18:10:05.166620 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-mvjvw_045426c7-6605-4623-93c3-1575773a5ecc/cert-manager-controller/0.log" Jan 21 18:10:05 crc kubenswrapper[4808]: I0121 18:10:05.180789 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-8jch4_71092cb7-08bc-4f0f-9696-838e2fc52d62/cert-manager-cainjector/0.log" Jan 21 18:10:05 crc kubenswrapper[4808]: I0121 18:10:05.193731 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-db8jl_d30e2538-925e-4c64-8d47-7d0b7255f69d/cert-manager-webhook/0.log" Jan 21 18:10:05 crc kubenswrapper[4808]: I0121 18:10:05.661640 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-mvjvw_045426c7-6605-4623-93c3-1575773a5ecc/cert-manager-controller/0.log" Jan 21 18:10:05 crc kubenswrapper[4808]: I0121 18:10:05.670514 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-8jch4_71092cb7-08bc-4f0f-9696-838e2fc52d62/cert-manager-cainjector/0.log" Jan 21 18:10:05 crc kubenswrapper[4808]: I0121 18:10:05.678786 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-db8jl_d30e2538-925e-4c64-8d47-7d0b7255f69d/cert-manager-webhook/0.log" Jan 21 18:10:06 crc kubenswrapper[4808]: I0121 18:10:06.092639 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-jl78z_11782075-8384-415d-a1b9-6c972f8f6f66/control-plane-machine-set-operator/0.log" Jan 21 18:10:06 crc kubenswrapper[4808]: I0121 18:10:06.102082 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4g7bd_4c87e4f3-6255-4fe3-a52c-726a1acf7ba8/kube-rbac-proxy/0.log" Jan 21 18:10:06 crc kubenswrapper[4808]: I0121 18:10:06.109835 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4g7bd_4c87e4f3-6255-4fe3-a52c-726a1acf7ba8/machine-api-operator/0.log" Jan 21 18:10:06 crc kubenswrapper[4808]: I0121 18:10:06.514377 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elastic-operator-7bddffb8f9-vlbqx_2ce84ba2-aeb7-40c7-bcb0-09aeb792b5bc/manager/0.log" Jan 21 18:10:06 crc kubenswrapper[4808]: I0121 18:10:06.538696 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elasticsearch-es-default-0_a879c359-bceb-47c5-a3db-e41219bc698d/elasticsearch/0.log" Jan 21 18:10:06 crc kubenswrapper[4808]: I0121 18:10:06.545750 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elasticsearch-es-default-0_a879c359-bceb-47c5-a3db-e41219bc698d/elastic-internal-init-filesystem/0.log" Jan 21 18:10:06 crc kubenswrapper[4808]: I0121 18:10:06.552341 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elasticsearch-es-default-0_a879c359-bceb-47c5-a3db-e41219bc698d/elastic-internal-suspend/0.log" Jan 21 18:10:06 crc kubenswrapper[4808]: I0121 18:10:06.563969 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_interconnect-operator-5bb49f789d-g47xc_47dc84b9-86ae-4fbb-af34-c628708ac495/interconnect-operator/0.log" Jan 21 18:10:07 crc kubenswrapper[4808]: I0121 18:10:07.306473 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-4kc58_4ffed385-f2d1-4043-9323-561451d85fad/kube-multus-additional-cni-plugins/0.log" Jan 21 18:10:07 crc kubenswrapper[4808]: I0121 18:10:07.323037 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-4kc58_4ffed385-f2d1-4043-9323-561451d85fad/egress-router-binary-copy/0.log" Jan 21 18:10:07 crc kubenswrapper[4808]: I0121 18:10:07.332047 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-4kc58_4ffed385-f2d1-4043-9323-561451d85fad/cni-plugins/0.log" Jan 21 18:10:07 crc kubenswrapper[4808]: I0121 18:10:07.339312 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-4kc58_4ffed385-f2d1-4043-9323-561451d85fad/bond-cni-plugin/0.log" Jan 21 18:10:07 crc kubenswrapper[4808]: I0121 18:10:07.346831 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-4kc58_4ffed385-f2d1-4043-9323-561451d85fad/routeoverride-cni/0.log" Jan 21 18:10:07 crc kubenswrapper[4808]: I0121 18:10:07.355156 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-4kc58_4ffed385-f2d1-4043-9323-561451d85fad/whereabouts-cni-bincopy/0.log" Jan 21 18:10:07 crc kubenswrapper[4808]: I0121 18:10:07.364513 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-4kc58_4ffed385-f2d1-4043-9323-561451d85fad/whereabouts-cni/0.log" Jan 21 18:10:07 crc kubenswrapper[4808]: I0121 18:10:07.375994 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-857f4d67dd-jxgs8_b5ad8de7-53f2-475d-be7b-acb85ad7f53a/multus-admission-controller/0.log" Jan 21 18:10:07 crc kubenswrapper[4808]: I0121 18:10:07.384752 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-857f4d67dd-jxgs8_b5ad8de7-53f2-475d-be7b-acb85ad7f53a/kube-rbac-proxy/0.log" Jan 21 18:10:07 crc kubenswrapper[4808]: I0121 18:10:07.408473 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8fp9_538d1d44-6c67-450a-809c-d9008128ec9f/kube-multus/3.log" Jan 21 18:10:07 crc kubenswrapper[4808]: I0121 18:10:07.428212 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8fp9_538d1d44-6c67-450a-809c-d9008128ec9f/kube-multus/2.log" Jan 21 18:10:07 crc kubenswrapper[4808]: I0121 18:10:07.447220 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-gm2t2_8d39c565-f404-451e-896f-ecc5af76a000/network-metrics-daemon/0.log" Jan 21 18:10:07 crc kubenswrapper[4808]: I0121 18:10:07.453013 4808 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-gm2t2_8d39c565-f404-451e-896f-ecc5af76a000/kube-rbac-proxy/0.log" Jan 21 18:10:42 crc kubenswrapper[4808]: I0121 18:10:42.304124 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-97kvw"] Jan 21 18:10:42 crc kubenswrapper[4808]: I0121 18:10:42.308634 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-97kvw" Jan 21 18:10:42 crc kubenswrapper[4808]: I0121 18:10:42.320266 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-97kvw"] Jan 21 18:10:42 crc kubenswrapper[4808]: I0121 18:10:42.437301 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll8p8\" (UniqueName: \"kubernetes.io/projected/9affcfe2-bcd3-44ef-ad84-24978a483620-kube-api-access-ll8p8\") pod \"redhat-operators-97kvw\" (UID: \"9affcfe2-bcd3-44ef-ad84-24978a483620\") " pod="openshift-marketplace/redhat-operators-97kvw" Jan 21 18:10:42 crc kubenswrapper[4808]: I0121 18:10:42.437397 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9affcfe2-bcd3-44ef-ad84-24978a483620-utilities\") pod \"redhat-operators-97kvw\" (UID: \"9affcfe2-bcd3-44ef-ad84-24978a483620\") " pod="openshift-marketplace/redhat-operators-97kvw" Jan 21 18:10:42 crc kubenswrapper[4808]: I0121 18:10:42.437589 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9affcfe2-bcd3-44ef-ad84-24978a483620-catalog-content\") pod \"redhat-operators-97kvw\" (UID: \"9affcfe2-bcd3-44ef-ad84-24978a483620\") " pod="openshift-marketplace/redhat-operators-97kvw" Jan 21 18:10:42 crc kubenswrapper[4808]: I0121 18:10:42.538873 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll8p8\" (UniqueName: \"kubernetes.io/projected/9affcfe2-bcd3-44ef-ad84-24978a483620-kube-api-access-ll8p8\") pod \"redhat-operators-97kvw\" (UID: \"9affcfe2-bcd3-44ef-ad84-24978a483620\") " pod="openshift-marketplace/redhat-operators-97kvw" Jan 21 18:10:42 crc kubenswrapper[4808]: I0121 18:10:42.538943 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9affcfe2-bcd3-44ef-ad84-24978a483620-utilities\") pod \"redhat-operators-97kvw\" (UID: \"9affcfe2-bcd3-44ef-ad84-24978a483620\") " pod="openshift-marketplace/redhat-operators-97kvw" Jan 21 18:10:42 crc kubenswrapper[4808]: I0121 18:10:42.538983 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9affcfe2-bcd3-44ef-ad84-24978a483620-catalog-content\") pod \"redhat-operators-97kvw\" (UID: \"9affcfe2-bcd3-44ef-ad84-24978a483620\") " pod="openshift-marketplace/redhat-operators-97kvw" Jan 21 18:10:42 crc kubenswrapper[4808]: I0121 18:10:42.539550 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9affcfe2-bcd3-44ef-ad84-24978a483620-catalog-content\") pod \"redhat-operators-97kvw\" (UID: \"9affcfe2-bcd3-44ef-ad84-24978a483620\") " pod="openshift-marketplace/redhat-operators-97kvw" Jan 21 18:10:42 crc kubenswrapper[4808]: I0121 18:10:42.539763 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9affcfe2-bcd3-44ef-ad84-24978a483620-utilities\") pod \"redhat-operators-97kvw\" (UID: \"9affcfe2-bcd3-44ef-ad84-24978a483620\") " pod="openshift-marketplace/redhat-operators-97kvw" Jan 21 18:10:42 crc kubenswrapper[4808]: I0121 18:10:42.558685 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll8p8\" (UniqueName: \"kubernetes.io/projected/9affcfe2-bcd3-44ef-ad84-24978a483620-kube-api-access-ll8p8\") pod \"redhat-operators-97kvw\" (UID: \"9affcfe2-bcd3-44ef-ad84-24978a483620\") " pod="openshift-marketplace/redhat-operators-97kvw" Jan 21 18:10:42 crc kubenswrapper[4808]: I0121 18:10:42.670120 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-97kvw" Jan 21 18:10:42 crc kubenswrapper[4808]: I0121 18:10:42.915200 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-97kvw"] Jan 21 18:10:43 crc kubenswrapper[4808]: I0121 18:10:43.198923 4808 generic.go:334] "Generic (PLEG): container finished" podID="9affcfe2-bcd3-44ef-ad84-24978a483620" containerID="5fcc49893303be815d1d8ced4254851e6f48231626cce5b8b9b9747999d396ce" exitCode=0 Jan 21 18:10:43 crc kubenswrapper[4808]: I0121 18:10:43.198983 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-97kvw" event={"ID":"9affcfe2-bcd3-44ef-ad84-24978a483620","Type":"ContainerDied","Data":"5fcc49893303be815d1d8ced4254851e6f48231626cce5b8b9b9747999d396ce"} Jan 21 18:10:43 crc kubenswrapper[4808]: I0121 18:10:43.199049 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-97kvw" event={"ID":"9affcfe2-bcd3-44ef-ad84-24978a483620","Type":"ContainerStarted","Data":"f7880e07d4f357ace4d0e9ed57daace461ef015d6f8ddfadc2ddc36f5e898138"} Jan 21 18:10:44 crc kubenswrapper[4808]: I0121 18:10:44.209605 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-97kvw" event={"ID":"9affcfe2-bcd3-44ef-ad84-24978a483620","Type":"ContainerStarted","Data":"164b6cdd9a180920f5d8685d5f0f0df98fcbda6829a9f6267fab4e8b53cd2116"} Jan 21 18:10:45 crc kubenswrapper[4808]: I0121 18:10:45.086393 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gmjzj"] Jan 21 18:10:45 crc kubenswrapper[4808]: I0121 18:10:45.091595 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gmjzj" Jan 21 18:10:45 crc kubenswrapper[4808]: I0121 18:10:45.097406 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gmjzj"] Jan 21 18:10:45 crc kubenswrapper[4808]: I0121 18:10:45.176309 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n8d6\" (UniqueName: \"kubernetes.io/projected/25818535-6cfc-4cf2-abf5-be6e9408e9ad-kube-api-access-9n8d6\") pod \"certified-operators-gmjzj\" (UID: \"25818535-6cfc-4cf2-abf5-be6e9408e9ad\") " pod="openshift-marketplace/certified-operators-gmjzj" Jan 21 18:10:45 crc kubenswrapper[4808]: I0121 18:10:45.176361 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25818535-6cfc-4cf2-abf5-be6e9408e9ad-catalog-content\") pod \"certified-operators-gmjzj\" (UID: \"25818535-6cfc-4cf2-abf5-be6e9408e9ad\") " pod="openshift-marketplace/certified-operators-gmjzj" Jan 21 18:10:45 crc kubenswrapper[4808]: I0121 18:10:45.176468 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25818535-6cfc-4cf2-abf5-be6e9408e9ad-utilities\") pod \"certified-operators-gmjzj\" (UID: \"25818535-6cfc-4cf2-abf5-be6e9408e9ad\") " pod="openshift-marketplace/certified-operators-gmjzj" Jan 21 18:10:45 crc kubenswrapper[4808]: I0121 18:10:45.217325 4808 generic.go:334] "Generic (PLEG): container finished" podID="9affcfe2-bcd3-44ef-ad84-24978a483620" containerID="164b6cdd9a180920f5d8685d5f0f0df98fcbda6829a9f6267fab4e8b53cd2116" exitCode=0 Jan 21 18:10:45 crc kubenswrapper[4808]: I0121 18:10:45.217367 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-97kvw" event={"ID":"9affcfe2-bcd3-44ef-ad84-24978a483620","Type":"ContainerDied","Data":"164b6cdd9a180920f5d8685d5f0f0df98fcbda6829a9f6267fab4e8b53cd2116"} Jan 21 18:10:45 crc kubenswrapper[4808]: I0121 18:10:45.277854 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n8d6\" (UniqueName: \"kubernetes.io/projected/25818535-6cfc-4cf2-abf5-be6e9408e9ad-kube-api-access-9n8d6\") pod \"certified-operators-gmjzj\" (UID: \"25818535-6cfc-4cf2-abf5-be6e9408e9ad\") " pod="openshift-marketplace/certified-operators-gmjzj" Jan 21 18:10:45 crc kubenswrapper[4808]: I0121 18:10:45.278268 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25818535-6cfc-4cf2-abf5-be6e9408e9ad-catalog-content\") pod \"certified-operators-gmjzj\" (UID: \"25818535-6cfc-4cf2-abf5-be6e9408e9ad\") " pod="openshift-marketplace/certified-operators-gmjzj" Jan 21 18:10:45 crc kubenswrapper[4808]: I0121 18:10:45.278355 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25818535-6cfc-4cf2-abf5-be6e9408e9ad-utilities\") pod \"certified-operators-gmjzj\" (UID: \"25818535-6cfc-4cf2-abf5-be6e9408e9ad\") " pod="openshift-marketplace/certified-operators-gmjzj" Jan 21 18:10:45 crc kubenswrapper[4808]: I0121 18:10:45.278724 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25818535-6cfc-4cf2-abf5-be6e9408e9ad-catalog-content\") pod \"certified-operators-gmjzj\" (UID: \"25818535-6cfc-4cf2-abf5-be6e9408e9ad\") " pod="openshift-marketplace/certified-operators-gmjzj" Jan 21 18:10:45 crc kubenswrapper[4808]: I0121 18:10:45.278808 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25818535-6cfc-4cf2-abf5-be6e9408e9ad-utilities\") pod \"certified-operators-gmjzj\" (UID: \"25818535-6cfc-4cf2-abf5-be6e9408e9ad\") " pod="openshift-marketplace/certified-operators-gmjzj" Jan 21 18:10:45 crc kubenswrapper[4808]: I0121 18:10:45.315232 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n8d6\" (UniqueName: \"kubernetes.io/projected/25818535-6cfc-4cf2-abf5-be6e9408e9ad-kube-api-access-9n8d6\") pod \"certified-operators-gmjzj\" (UID: \"25818535-6cfc-4cf2-abf5-be6e9408e9ad\") " pod="openshift-marketplace/certified-operators-gmjzj" Jan 21 18:10:45 crc kubenswrapper[4808]: I0121 18:10:45.416379 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gmjzj" Jan 21 18:10:45 crc kubenswrapper[4808]: I0121 18:10:45.680257 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gmjzj"] Jan 21 18:10:46 crc kubenswrapper[4808]: I0121 18:10:46.227728 4808 generic.go:334] "Generic (PLEG): container finished" podID="25818535-6cfc-4cf2-abf5-be6e9408e9ad" containerID="4eed8de69d939111e9cfc9a19fa82f46b5c97c1cc6c1b311e030060bb1171445" exitCode=0 Jan 21 18:10:46 crc kubenswrapper[4808]: I0121 18:10:46.227797 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmjzj" event={"ID":"25818535-6cfc-4cf2-abf5-be6e9408e9ad","Type":"ContainerDied","Data":"4eed8de69d939111e9cfc9a19fa82f46b5c97c1cc6c1b311e030060bb1171445"} Jan 21 18:10:46 crc kubenswrapper[4808]: I0121 18:10:46.228165 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmjzj" event={"ID":"25818535-6cfc-4cf2-abf5-be6e9408e9ad","Type":"ContainerStarted","Data":"d1cfbb361286b52f09e56e389fcabc978052da286a700c97142eb0ba81cc7680"} Jan 21 18:10:46 crc kubenswrapper[4808]: I0121 18:10:46.232367 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-97kvw" event={"ID":"9affcfe2-bcd3-44ef-ad84-24978a483620","Type":"ContainerStarted","Data":"c5702133ab581e856e92c4c4df8bde2199c9785d318b359e7f8d3e27cac3ca69"} Jan 21 18:10:46 crc kubenswrapper[4808]: I0121 18:10:46.290004 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-97kvw" podStartSLOduration=1.658076605 podStartE2EDuration="4.289983597s" podCreationTimestamp="2026-01-21 18:10:42 +0000 UTC" firstStartedPulling="2026-01-21 18:10:43.200931101 +0000 UTC m=+877.931441986" lastFinishedPulling="2026-01-21 18:10:45.832838093 +0000 UTC m=+880.563348978" observedRunningTime="2026-01-21 18:10:46.287042462 +0000 UTC m=+881.017553407" watchObservedRunningTime="2026-01-21 18:10:46.289983597 +0000 UTC m=+881.020494492" Jan 21 18:10:48 crc kubenswrapper[4808]: I0121 18:10:48.249693 4808 generic.go:334] "Generic (PLEG): container finished" podID="25818535-6cfc-4cf2-abf5-be6e9408e9ad" containerID="0b25332002b0cabd662366e51d11d1dc7166a20a0e66b74cb167e172d4a26778" exitCode=0 Jan 21 18:10:48 crc kubenswrapper[4808]: I0121 18:10:48.249752 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmjzj" event={"ID":"25818535-6cfc-4cf2-abf5-be6e9408e9ad","Type":"ContainerDied","Data":"0b25332002b0cabd662366e51d11d1dc7166a20a0e66b74cb167e172d4a26778"} Jan 21 18:10:51 crc kubenswrapper[4808]: I0121 18:10:51.272801 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmjzj" event={"ID":"25818535-6cfc-4cf2-abf5-be6e9408e9ad","Type":"ContainerStarted","Data":"5706cd15bb75b56a5465b21e00108afb8f049d297762445b415d19038b8a1394"} Jan 21 18:10:51 crc kubenswrapper[4808]: I0121 18:10:51.292984 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gmjzj" podStartSLOduration=2.410046861 podStartE2EDuration="6.292962684s" podCreationTimestamp="2026-01-21 18:10:45 +0000 UTC" firstStartedPulling="2026-01-21 18:10:46.229215983 +0000 UTC m=+880.959726908" lastFinishedPulling="2026-01-21 18:10:50.112131836 +0000 UTC m=+884.842642731" observedRunningTime="2026-01-21 18:10:51.291088446 +0000 UTC m=+886.021599341" watchObservedRunningTime="2026-01-21 18:10:51.292962684 +0000 UTC m=+886.023473569" Jan 21 18:10:52 crc kubenswrapper[4808]: I0121 18:10:52.285798 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vhqg2"] Jan 21 18:10:52 crc kubenswrapper[4808]: I0121 18:10:52.287859 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vhqg2" Jan 21 18:10:52 crc kubenswrapper[4808]: I0121 18:10:52.304327 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vhqg2"] Jan 21 18:10:52 crc kubenswrapper[4808]: I0121 18:10:52.377301 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9667435d-cc17-4d75-8a83-3fbbd87ef20f-utilities\") pod \"community-operators-vhqg2\" (UID: \"9667435d-cc17-4d75-8a83-3fbbd87ef20f\") " pod="openshift-marketplace/community-operators-vhqg2" Jan 21 18:10:52 crc kubenswrapper[4808]: I0121 18:10:52.377351 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkwhm\" (UniqueName: \"kubernetes.io/projected/9667435d-cc17-4d75-8a83-3fbbd87ef20f-kube-api-access-fkwhm\") pod \"community-operators-vhqg2\" (UID: \"9667435d-cc17-4d75-8a83-3fbbd87ef20f\") " pod="openshift-marketplace/community-operators-vhqg2" Jan 21 18:10:52 crc kubenswrapper[4808]: I0121 18:10:52.377401 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9667435d-cc17-4d75-8a83-3fbbd87ef20f-catalog-content\") pod \"community-operators-vhqg2\" (UID: \"9667435d-cc17-4d75-8a83-3fbbd87ef20f\") " pod="openshift-marketplace/community-operators-vhqg2" Jan 21 18:10:52 crc kubenswrapper[4808]: I0121 18:10:52.478470 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9667435d-cc17-4d75-8a83-3fbbd87ef20f-utilities\") pod \"community-operators-vhqg2\" (UID: \"9667435d-cc17-4d75-8a83-3fbbd87ef20f\") " pod="openshift-marketplace/community-operators-vhqg2" Jan 21 18:10:52 crc kubenswrapper[4808]: I0121 18:10:52.478524 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkwhm\" (UniqueName: \"kubernetes.io/projected/9667435d-cc17-4d75-8a83-3fbbd87ef20f-kube-api-access-fkwhm\") pod \"community-operators-vhqg2\" (UID: \"9667435d-cc17-4d75-8a83-3fbbd87ef20f\") " pod="openshift-marketplace/community-operators-vhqg2" Jan 21 18:10:52 crc kubenswrapper[4808]: I0121 18:10:52.478581 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9667435d-cc17-4d75-8a83-3fbbd87ef20f-catalog-content\") pod \"community-operators-vhqg2\" (UID: \"9667435d-cc17-4d75-8a83-3fbbd87ef20f\") " pod="openshift-marketplace/community-operators-vhqg2" Jan 21 18:10:52 crc kubenswrapper[4808]: I0121 18:10:52.479024 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9667435d-cc17-4d75-8a83-3fbbd87ef20f-utilities\") pod \"community-operators-vhqg2\" (UID: \"9667435d-cc17-4d75-8a83-3fbbd87ef20f\") " pod="openshift-marketplace/community-operators-vhqg2" Jan 21 18:10:52 crc kubenswrapper[4808]: I0121 18:10:52.479051 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9667435d-cc17-4d75-8a83-3fbbd87ef20f-catalog-content\") pod \"community-operators-vhqg2\" (UID: \"9667435d-cc17-4d75-8a83-3fbbd87ef20f\") " pod="openshift-marketplace/community-operators-vhqg2" Jan 21 18:10:52 crc kubenswrapper[4808]: I0121 18:10:52.497466 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkwhm\" (UniqueName: \"kubernetes.io/projected/9667435d-cc17-4d75-8a83-3fbbd87ef20f-kube-api-access-fkwhm\") pod \"community-operators-vhqg2\" (UID: \"9667435d-cc17-4d75-8a83-3fbbd87ef20f\") " pod="openshift-marketplace/community-operators-vhqg2" Jan 21 18:10:52 crc kubenswrapper[4808]: I0121 18:10:52.621514 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vhqg2" Jan 21 18:10:52 crc kubenswrapper[4808]: I0121 18:10:52.671194 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-97kvw" Jan 21 18:10:52 crc kubenswrapper[4808]: I0121 18:10:52.671264 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-97kvw" Jan 21 18:10:52 crc kubenswrapper[4808]: I0121 18:10:52.725014 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-97kvw" Jan 21 18:10:52 crc kubenswrapper[4808]: I0121 18:10:52.917517 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vhqg2"] Jan 21 18:10:53 crc kubenswrapper[4808]: I0121 18:10:53.295481 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vhqg2" event={"ID":"9667435d-cc17-4d75-8a83-3fbbd87ef20f","Type":"ContainerStarted","Data":"21250bbf881927a048dad8a97c51c929f2ecb95507e06650ea92a84f9937aeb4"} Jan 21 18:10:53 crc kubenswrapper[4808]: I0121 18:10:53.295566 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vhqg2" event={"ID":"9667435d-cc17-4d75-8a83-3fbbd87ef20f","Type":"ContainerStarted","Data":"98d764846755f31c383f03f5e622e989bb3d429c948714dd40b515039b019d66"} Jan 21 18:10:53 crc kubenswrapper[4808]: I0121 18:10:53.377273 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-97kvw" Jan 21 18:10:54 crc kubenswrapper[4808]: I0121 18:10:54.303140 4808 generic.go:334] "Generic (PLEG): container finished" podID="9667435d-cc17-4d75-8a83-3fbbd87ef20f" containerID="21250bbf881927a048dad8a97c51c929f2ecb95507e06650ea92a84f9937aeb4" exitCode=0 Jan 21 18:10:54 crc kubenswrapper[4808]: I0121 18:10:54.303371 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vhqg2" event={"ID":"9667435d-cc17-4d75-8a83-3fbbd87ef20f","Type":"ContainerDied","Data":"21250bbf881927a048dad8a97c51c929f2ecb95507e06650ea92a84f9937aeb4"} Jan 21 18:10:55 crc kubenswrapper[4808]: I0121 18:10:55.416511 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gmjzj" Jan 21 18:10:55 crc kubenswrapper[4808]: I0121 18:10:55.417440 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gmjzj" Jan 21 18:10:55 crc kubenswrapper[4808]: I0121 18:10:55.471666 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gmjzj" Jan 21 18:10:55 crc kubenswrapper[4808]: I0121 18:10:55.674807 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-97kvw"] Jan 21 18:10:55 crc kubenswrapper[4808]: I0121 18:10:55.675023 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-97kvw" podUID="9affcfe2-bcd3-44ef-ad84-24978a483620" containerName="registry-server" containerID="cri-o://c5702133ab581e856e92c4c4df8bde2199c9785d318b359e7f8d3e27cac3ca69" gracePeriod=2 Jan 21 18:10:56 crc kubenswrapper[4808]: I0121 18:10:56.361812 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gmjzj" Jan 21 18:10:57 crc kubenswrapper[4808]: I0121 18:10:57.840964 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-97kvw" Jan 21 18:10:57 crc kubenswrapper[4808]: I0121 18:10:57.961562 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ll8p8\" (UniqueName: \"kubernetes.io/projected/9affcfe2-bcd3-44ef-ad84-24978a483620-kube-api-access-ll8p8\") pod \"9affcfe2-bcd3-44ef-ad84-24978a483620\" (UID: \"9affcfe2-bcd3-44ef-ad84-24978a483620\") " Jan 21 18:10:57 crc kubenswrapper[4808]: I0121 18:10:57.961819 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9affcfe2-bcd3-44ef-ad84-24978a483620-catalog-content\") pod \"9affcfe2-bcd3-44ef-ad84-24978a483620\" (UID: \"9affcfe2-bcd3-44ef-ad84-24978a483620\") " Jan 21 18:10:57 crc kubenswrapper[4808]: I0121 18:10:57.961881 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9affcfe2-bcd3-44ef-ad84-24978a483620-utilities\") pod \"9affcfe2-bcd3-44ef-ad84-24978a483620\" (UID: \"9affcfe2-bcd3-44ef-ad84-24978a483620\") " Jan 21 18:10:57 crc kubenswrapper[4808]: I0121 18:10:57.962581 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9affcfe2-bcd3-44ef-ad84-24978a483620-utilities" (OuterVolumeSpecName: "utilities") pod "9affcfe2-bcd3-44ef-ad84-24978a483620" (UID: "9affcfe2-bcd3-44ef-ad84-24978a483620"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:10:57 crc kubenswrapper[4808]: I0121 18:10:57.968488 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9affcfe2-bcd3-44ef-ad84-24978a483620-kube-api-access-ll8p8" (OuterVolumeSpecName: "kube-api-access-ll8p8") pod "9affcfe2-bcd3-44ef-ad84-24978a483620" (UID: "9affcfe2-bcd3-44ef-ad84-24978a483620"). InnerVolumeSpecName "kube-api-access-ll8p8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.063536 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9affcfe2-bcd3-44ef-ad84-24978a483620-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.063589 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ll8p8\" (UniqueName: \"kubernetes.io/projected/9affcfe2-bcd3-44ef-ad84-24978a483620-kube-api-access-ll8p8\") on node \"crc\" DevicePath \"\"" Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.074083 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9affcfe2-bcd3-44ef-ad84-24978a483620-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9affcfe2-bcd3-44ef-ad84-24978a483620" (UID: "9affcfe2-bcd3-44ef-ad84-24978a483620"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.079979 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gmjzj"] Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.165216 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9affcfe2-bcd3-44ef-ad84-24978a483620-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.351615 4808 generic.go:334] "Generic (PLEG): container finished" podID="9affcfe2-bcd3-44ef-ad84-24978a483620" containerID="c5702133ab581e856e92c4c4df8bde2199c9785d318b359e7f8d3e27cac3ca69" exitCode=0 Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.351673 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-97kvw" event={"ID":"9affcfe2-bcd3-44ef-ad84-24978a483620","Type":"ContainerDied","Data":"c5702133ab581e856e92c4c4df8bde2199c9785d318b359e7f8d3e27cac3ca69"} Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.351697 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-97kvw" event={"ID":"9affcfe2-bcd3-44ef-ad84-24978a483620","Type":"ContainerDied","Data":"f7880e07d4f357ace4d0e9ed57daace461ef015d6f8ddfadc2ddc36f5e898138"} Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.351717 4808 scope.go:117] "RemoveContainer" containerID="c5702133ab581e856e92c4c4df8bde2199c9785d318b359e7f8d3e27cac3ca69" Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.351836 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-97kvw" Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.355703 4808 generic.go:334] "Generic (PLEG): container finished" podID="9667435d-cc17-4d75-8a83-3fbbd87ef20f" containerID="f3b99ed4c156769f1e9dbd9952e637c4d0ffe19173eba5f1879fd8d46a3d92b6" exitCode=0 Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.356866 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vhqg2" event={"ID":"9667435d-cc17-4d75-8a83-3fbbd87ef20f","Type":"ContainerDied","Data":"f3b99ed4c156769f1e9dbd9952e637c4d0ffe19173eba5f1879fd8d46a3d92b6"} Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.375297 4808 scope.go:117] "RemoveContainer" containerID="164b6cdd9a180920f5d8685d5f0f0df98fcbda6829a9f6267fab4e8b53cd2116" Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.399976 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-97kvw"] Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.406783 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-97kvw"] Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.408574 4808 scope.go:117] "RemoveContainer" containerID="5fcc49893303be815d1d8ced4254851e6f48231626cce5b8b9b9747999d396ce" Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.430282 4808 scope.go:117] "RemoveContainer" containerID="c5702133ab581e856e92c4c4df8bde2199c9785d318b359e7f8d3e27cac3ca69" Jan 21 18:10:58 crc kubenswrapper[4808]: E0121 18:10:58.430646 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5702133ab581e856e92c4c4df8bde2199c9785d318b359e7f8d3e27cac3ca69\": container with ID starting with c5702133ab581e856e92c4c4df8bde2199c9785d318b359e7f8d3e27cac3ca69 not found: ID does not exist" containerID="c5702133ab581e856e92c4c4df8bde2199c9785d318b359e7f8d3e27cac3ca69" Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.430686 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5702133ab581e856e92c4c4df8bde2199c9785d318b359e7f8d3e27cac3ca69"} err="failed to get container status \"c5702133ab581e856e92c4c4df8bde2199c9785d318b359e7f8d3e27cac3ca69\": rpc error: code = NotFound desc = could not find container \"c5702133ab581e856e92c4c4df8bde2199c9785d318b359e7f8d3e27cac3ca69\": container with ID starting with c5702133ab581e856e92c4c4df8bde2199c9785d318b359e7f8d3e27cac3ca69 not found: ID does not exist" Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.430712 4808 scope.go:117] "RemoveContainer" containerID="164b6cdd9a180920f5d8685d5f0f0df98fcbda6829a9f6267fab4e8b53cd2116" Jan 21 18:10:58 crc kubenswrapper[4808]: E0121 18:10:58.431098 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"164b6cdd9a180920f5d8685d5f0f0df98fcbda6829a9f6267fab4e8b53cd2116\": container with ID starting with 164b6cdd9a180920f5d8685d5f0f0df98fcbda6829a9f6267fab4e8b53cd2116 not found: ID does not exist" containerID="164b6cdd9a180920f5d8685d5f0f0df98fcbda6829a9f6267fab4e8b53cd2116" Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.431126 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"164b6cdd9a180920f5d8685d5f0f0df98fcbda6829a9f6267fab4e8b53cd2116"} err="failed to get container status \"164b6cdd9a180920f5d8685d5f0f0df98fcbda6829a9f6267fab4e8b53cd2116\": rpc error: code = NotFound desc = could not find container \"164b6cdd9a180920f5d8685d5f0f0df98fcbda6829a9f6267fab4e8b53cd2116\": container with ID starting with 164b6cdd9a180920f5d8685d5f0f0df98fcbda6829a9f6267fab4e8b53cd2116 not found: ID does not exist" Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.431143 4808 scope.go:117] "RemoveContainer" containerID="5fcc49893303be815d1d8ced4254851e6f48231626cce5b8b9b9747999d396ce" Jan 21 18:10:58 crc kubenswrapper[4808]: E0121 18:10:58.431405 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fcc49893303be815d1d8ced4254851e6f48231626cce5b8b9b9747999d396ce\": container with ID starting with 5fcc49893303be815d1d8ced4254851e6f48231626cce5b8b9b9747999d396ce not found: ID does not exist" containerID="5fcc49893303be815d1d8ced4254851e6f48231626cce5b8b9b9747999d396ce" Jan 21 18:10:58 crc kubenswrapper[4808]: I0121 18:10:58.431434 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fcc49893303be815d1d8ced4254851e6f48231626cce5b8b9b9747999d396ce"} err="failed to get container status \"5fcc49893303be815d1d8ced4254851e6f48231626cce5b8b9b9747999d396ce\": rpc error: code = NotFound desc = could not find container \"5fcc49893303be815d1d8ced4254851e6f48231626cce5b8b9b9747999d396ce\": container with ID starting with 5fcc49893303be815d1d8ced4254851e6f48231626cce5b8b9b9747999d396ce not found: ID does not exist" Jan 21 18:10:58 crc kubenswrapper[4808]: E0121 18:10:58.436101 4808 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9affcfe2_bcd3_44ef_ad84_24978a483620.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9affcfe2_bcd3_44ef_ad84_24978a483620.slice/crio-f7880e07d4f357ace4d0e9ed57daace461ef015d6f8ddfadc2ddc36f5e898138\": RecentStats: unable to find data in memory cache]" Jan 21 18:10:59 crc kubenswrapper[4808]: I0121 18:10:59.365136 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vhqg2" event={"ID":"9667435d-cc17-4d75-8a83-3fbbd87ef20f","Type":"ContainerStarted","Data":"6abf2cdf457bf50e514b7a5c1646abd69c41b889cacaa4b6022179ffcc105e2c"} Jan 21 18:10:59 crc kubenswrapper[4808]: I0121 18:10:59.367347 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gmjzj" podUID="25818535-6cfc-4cf2-abf5-be6e9408e9ad" containerName="registry-server" containerID="cri-o://5706cd15bb75b56a5465b21e00108afb8f049d297762445b415d19038b8a1394" gracePeriod=2 Jan 21 18:10:59 crc kubenswrapper[4808]: I0121 18:10:59.391538 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vhqg2" podStartSLOduration=2.944193941 podStartE2EDuration="7.391522013s" podCreationTimestamp="2026-01-21 18:10:52 +0000 UTC" firstStartedPulling="2026-01-21 18:10:54.305628799 +0000 UTC m=+889.036139684" lastFinishedPulling="2026-01-21 18:10:58.752956871 +0000 UTC m=+893.483467756" observedRunningTime="2026-01-21 18:10:59.389399699 +0000 UTC m=+894.119910574" watchObservedRunningTime="2026-01-21 18:10:59.391522013 +0000 UTC m=+894.122032898" Jan 21 18:10:59 crc kubenswrapper[4808]: I0121 18:10:59.529580 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9affcfe2-bcd3-44ef-ad84-24978a483620" path="/var/lib/kubelet/pods/9affcfe2-bcd3-44ef-ad84-24978a483620/volumes" Jan 21 18:10:59 crc kubenswrapper[4808]: I0121 18:10:59.714058 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gmjzj" Jan 21 18:10:59 crc kubenswrapper[4808]: I0121 18:10:59.790935 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25818535-6cfc-4cf2-abf5-be6e9408e9ad-catalog-content\") pod \"25818535-6cfc-4cf2-abf5-be6e9408e9ad\" (UID: \"25818535-6cfc-4cf2-abf5-be6e9408e9ad\") " Jan 21 18:10:59 crc kubenswrapper[4808]: I0121 18:10:59.790992 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n8d6\" (UniqueName: \"kubernetes.io/projected/25818535-6cfc-4cf2-abf5-be6e9408e9ad-kube-api-access-9n8d6\") pod \"25818535-6cfc-4cf2-abf5-be6e9408e9ad\" (UID: \"25818535-6cfc-4cf2-abf5-be6e9408e9ad\") " Jan 21 18:10:59 crc kubenswrapper[4808]: I0121 18:10:59.791119 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25818535-6cfc-4cf2-abf5-be6e9408e9ad-utilities\") pod \"25818535-6cfc-4cf2-abf5-be6e9408e9ad\" (UID: \"25818535-6cfc-4cf2-abf5-be6e9408e9ad\") " Jan 21 18:10:59 crc kubenswrapper[4808]: I0121 18:10:59.791773 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25818535-6cfc-4cf2-abf5-be6e9408e9ad-utilities" (OuterVolumeSpecName: "utilities") pod "25818535-6cfc-4cf2-abf5-be6e9408e9ad" (UID: "25818535-6cfc-4cf2-abf5-be6e9408e9ad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:10:59 crc kubenswrapper[4808]: I0121 18:10:59.792108 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25818535-6cfc-4cf2-abf5-be6e9408e9ad-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:10:59 crc kubenswrapper[4808]: I0121 18:10:59.797385 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25818535-6cfc-4cf2-abf5-be6e9408e9ad-kube-api-access-9n8d6" (OuterVolumeSpecName: "kube-api-access-9n8d6") pod "25818535-6cfc-4cf2-abf5-be6e9408e9ad" (UID: "25818535-6cfc-4cf2-abf5-be6e9408e9ad"). InnerVolumeSpecName "kube-api-access-9n8d6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:10:59 crc kubenswrapper[4808]: I0121 18:10:59.840733 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25818535-6cfc-4cf2-abf5-be6e9408e9ad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25818535-6cfc-4cf2-abf5-be6e9408e9ad" (UID: "25818535-6cfc-4cf2-abf5-be6e9408e9ad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:10:59 crc kubenswrapper[4808]: I0121 18:10:59.893466 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25818535-6cfc-4cf2-abf5-be6e9408e9ad-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:10:59 crc kubenswrapper[4808]: I0121 18:10:59.893503 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n8d6\" (UniqueName: \"kubernetes.io/projected/25818535-6cfc-4cf2-abf5-be6e9408e9ad-kube-api-access-9n8d6\") on node \"crc\" DevicePath \"\"" Jan 21 18:11:00 crc kubenswrapper[4808]: I0121 18:11:00.376212 4808 generic.go:334] "Generic (PLEG): container finished" podID="25818535-6cfc-4cf2-abf5-be6e9408e9ad" containerID="5706cd15bb75b56a5465b21e00108afb8f049d297762445b415d19038b8a1394" exitCode=0 Jan 21 18:11:00 crc kubenswrapper[4808]: I0121 18:11:00.376637 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gmjzj" Jan 21 18:11:00 crc kubenswrapper[4808]: I0121 18:11:00.376962 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmjzj" event={"ID":"25818535-6cfc-4cf2-abf5-be6e9408e9ad","Type":"ContainerDied","Data":"5706cd15bb75b56a5465b21e00108afb8f049d297762445b415d19038b8a1394"} Jan 21 18:11:00 crc kubenswrapper[4808]: I0121 18:11:00.376999 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmjzj" event={"ID":"25818535-6cfc-4cf2-abf5-be6e9408e9ad","Type":"ContainerDied","Data":"d1cfbb361286b52f09e56e389fcabc978052da286a700c97142eb0ba81cc7680"} Jan 21 18:11:00 crc kubenswrapper[4808]: I0121 18:11:00.377021 4808 scope.go:117] "RemoveContainer" containerID="5706cd15bb75b56a5465b21e00108afb8f049d297762445b415d19038b8a1394" Jan 21 18:11:00 crc kubenswrapper[4808]: I0121 18:11:00.401344 4808 scope.go:117] "RemoveContainer" containerID="0b25332002b0cabd662366e51d11d1dc7166a20a0e66b74cb167e172d4a26778" Jan 21 18:11:00 crc kubenswrapper[4808]: I0121 18:11:00.411546 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gmjzj"] Jan 21 18:11:00 crc kubenswrapper[4808]: I0121 18:11:00.417510 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gmjzj"] Jan 21 18:11:00 crc kubenswrapper[4808]: I0121 18:11:00.445529 4808 scope.go:117] "RemoveContainer" containerID="4eed8de69d939111e9cfc9a19fa82f46b5c97c1cc6c1b311e030060bb1171445" Jan 21 18:11:00 crc kubenswrapper[4808]: I0121 18:11:00.465289 4808 scope.go:117] "RemoveContainer" containerID="5706cd15bb75b56a5465b21e00108afb8f049d297762445b415d19038b8a1394" Jan 21 18:11:00 crc kubenswrapper[4808]: E0121 18:11:00.465846 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5706cd15bb75b56a5465b21e00108afb8f049d297762445b415d19038b8a1394\": container with ID starting with 5706cd15bb75b56a5465b21e00108afb8f049d297762445b415d19038b8a1394 not found: ID does not exist" containerID="5706cd15bb75b56a5465b21e00108afb8f049d297762445b415d19038b8a1394" Jan 21 18:11:00 crc kubenswrapper[4808]: I0121 18:11:00.465873 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5706cd15bb75b56a5465b21e00108afb8f049d297762445b415d19038b8a1394"} err="failed to get container status \"5706cd15bb75b56a5465b21e00108afb8f049d297762445b415d19038b8a1394\": rpc error: code = NotFound desc = could not find container \"5706cd15bb75b56a5465b21e00108afb8f049d297762445b415d19038b8a1394\": container with ID starting with 5706cd15bb75b56a5465b21e00108afb8f049d297762445b415d19038b8a1394 not found: ID does not exist" Jan 21 18:11:00 crc kubenswrapper[4808]: I0121 18:11:00.465894 4808 scope.go:117] "RemoveContainer" containerID="0b25332002b0cabd662366e51d11d1dc7166a20a0e66b74cb167e172d4a26778" Jan 21 18:11:00 crc kubenswrapper[4808]: E0121 18:11:00.466231 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b25332002b0cabd662366e51d11d1dc7166a20a0e66b74cb167e172d4a26778\": container with ID starting with 0b25332002b0cabd662366e51d11d1dc7166a20a0e66b74cb167e172d4a26778 not found: ID does not exist" containerID="0b25332002b0cabd662366e51d11d1dc7166a20a0e66b74cb167e172d4a26778" Jan 21 18:11:00 crc kubenswrapper[4808]: I0121 18:11:00.466268 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b25332002b0cabd662366e51d11d1dc7166a20a0e66b74cb167e172d4a26778"} err="failed to get container status \"0b25332002b0cabd662366e51d11d1dc7166a20a0e66b74cb167e172d4a26778\": rpc error: code = NotFound desc = could not find container \"0b25332002b0cabd662366e51d11d1dc7166a20a0e66b74cb167e172d4a26778\": container with ID starting with 0b25332002b0cabd662366e51d11d1dc7166a20a0e66b74cb167e172d4a26778 not found: ID does not exist" Jan 21 18:11:00 crc kubenswrapper[4808]: I0121 18:11:00.466287 4808 scope.go:117] "RemoveContainer" containerID="4eed8de69d939111e9cfc9a19fa82f46b5c97c1cc6c1b311e030060bb1171445" Jan 21 18:11:00 crc kubenswrapper[4808]: E0121 18:11:00.466523 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4eed8de69d939111e9cfc9a19fa82f46b5c97c1cc6c1b311e030060bb1171445\": container with ID starting with 4eed8de69d939111e9cfc9a19fa82f46b5c97c1cc6c1b311e030060bb1171445 not found: ID does not exist" containerID="4eed8de69d939111e9cfc9a19fa82f46b5c97c1cc6c1b311e030060bb1171445" Jan 21 18:11:00 crc kubenswrapper[4808]: I0121 18:11:00.466542 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4eed8de69d939111e9cfc9a19fa82f46b5c97c1cc6c1b311e030060bb1171445"} err="failed to get container status \"4eed8de69d939111e9cfc9a19fa82f46b5c97c1cc6c1b311e030060bb1171445\": rpc error: code = NotFound desc = could not find container \"4eed8de69d939111e9cfc9a19fa82f46b5c97c1cc6c1b311e030060bb1171445\": container with ID starting with 4eed8de69d939111e9cfc9a19fa82f46b5c97c1cc6c1b311e030060bb1171445 not found: ID does not exist" Jan 21 18:11:01 crc kubenswrapper[4808]: I0121 18:11:01.541174 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25818535-6cfc-4cf2-abf5-be6e9408e9ad" path="/var/lib/kubelet/pods/25818535-6cfc-4cf2-abf5-be6e9408e9ad/volumes" Jan 21 18:11:02 crc kubenswrapper[4808]: I0121 18:11:02.622216 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vhqg2" Jan 21 18:11:02 crc kubenswrapper[4808]: I0121 18:11:02.623451 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vhqg2" Jan 21 18:11:02 crc kubenswrapper[4808]: I0121 18:11:02.658279 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vhqg2" Jan 21 18:11:03 crc kubenswrapper[4808]: I0121 18:11:03.447408 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vhqg2" Jan 21 18:11:04 crc kubenswrapper[4808]: I0121 18:11:04.470777 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vhqg2"] Jan 21 18:11:06 crc kubenswrapper[4808]: I0121 18:11:06.418580 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vhqg2" podUID="9667435d-cc17-4d75-8a83-3fbbd87ef20f" containerName="registry-server" containerID="cri-o://6abf2cdf457bf50e514b7a5c1646abd69c41b889cacaa4b6022179ffcc105e2c" gracePeriod=2 Jan 21 18:11:06 crc kubenswrapper[4808]: I0121 18:11:06.784107 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vhqg2" Jan 21 18:11:06 crc kubenswrapper[4808]: I0121 18:11:06.912824 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkwhm\" (UniqueName: \"kubernetes.io/projected/9667435d-cc17-4d75-8a83-3fbbd87ef20f-kube-api-access-fkwhm\") pod \"9667435d-cc17-4d75-8a83-3fbbd87ef20f\" (UID: \"9667435d-cc17-4d75-8a83-3fbbd87ef20f\") " Jan 21 18:11:06 crc kubenswrapper[4808]: I0121 18:11:06.912969 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9667435d-cc17-4d75-8a83-3fbbd87ef20f-utilities\") pod \"9667435d-cc17-4d75-8a83-3fbbd87ef20f\" (UID: \"9667435d-cc17-4d75-8a83-3fbbd87ef20f\") " Jan 21 18:11:06 crc kubenswrapper[4808]: I0121 18:11:06.913044 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9667435d-cc17-4d75-8a83-3fbbd87ef20f-catalog-content\") pod \"9667435d-cc17-4d75-8a83-3fbbd87ef20f\" (UID: \"9667435d-cc17-4d75-8a83-3fbbd87ef20f\") " Jan 21 18:11:06 crc kubenswrapper[4808]: I0121 18:11:06.916718 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9667435d-cc17-4d75-8a83-3fbbd87ef20f-utilities" (OuterVolumeSpecName: "utilities") pod "9667435d-cc17-4d75-8a83-3fbbd87ef20f" (UID: "9667435d-cc17-4d75-8a83-3fbbd87ef20f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:11:06 crc kubenswrapper[4808]: I0121 18:11:06.919909 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9667435d-cc17-4d75-8a83-3fbbd87ef20f-kube-api-access-fkwhm" (OuterVolumeSpecName: "kube-api-access-fkwhm") pod "9667435d-cc17-4d75-8a83-3fbbd87ef20f" (UID: "9667435d-cc17-4d75-8a83-3fbbd87ef20f"). InnerVolumeSpecName "kube-api-access-fkwhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:11:06 crc kubenswrapper[4808]: I0121 18:11:06.967852 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9667435d-cc17-4d75-8a83-3fbbd87ef20f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9667435d-cc17-4d75-8a83-3fbbd87ef20f" (UID: "9667435d-cc17-4d75-8a83-3fbbd87ef20f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:11:07 crc kubenswrapper[4808]: I0121 18:11:07.015304 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9667435d-cc17-4d75-8a83-3fbbd87ef20f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:11:07 crc kubenswrapper[4808]: I0121 18:11:07.015358 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkwhm\" (UniqueName: \"kubernetes.io/projected/9667435d-cc17-4d75-8a83-3fbbd87ef20f-kube-api-access-fkwhm\") on node \"crc\" DevicePath \"\"" Jan 21 18:11:07 crc kubenswrapper[4808]: I0121 18:11:07.015380 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9667435d-cc17-4d75-8a83-3fbbd87ef20f-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:11:07 crc kubenswrapper[4808]: I0121 18:11:07.427496 4808 generic.go:334] "Generic (PLEG): container finished" podID="9667435d-cc17-4d75-8a83-3fbbd87ef20f" containerID="6abf2cdf457bf50e514b7a5c1646abd69c41b889cacaa4b6022179ffcc105e2c" exitCode=0 Jan 21 18:11:07 crc kubenswrapper[4808]: I0121 18:11:07.427544 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vhqg2" event={"ID":"9667435d-cc17-4d75-8a83-3fbbd87ef20f","Type":"ContainerDied","Data":"6abf2cdf457bf50e514b7a5c1646abd69c41b889cacaa4b6022179ffcc105e2c"} Jan 21 18:11:07 crc kubenswrapper[4808]: I0121 18:11:07.427563 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vhqg2" Jan 21 18:11:07 crc kubenswrapper[4808]: I0121 18:11:07.427574 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vhqg2" event={"ID":"9667435d-cc17-4d75-8a83-3fbbd87ef20f","Type":"ContainerDied","Data":"98d764846755f31c383f03f5e622e989bb3d429c948714dd40b515039b019d66"} Jan 21 18:11:07 crc kubenswrapper[4808]: I0121 18:11:07.427596 4808 scope.go:117] "RemoveContainer" containerID="6abf2cdf457bf50e514b7a5c1646abd69c41b889cacaa4b6022179ffcc105e2c" Jan 21 18:11:07 crc kubenswrapper[4808]: I0121 18:11:07.450565 4808 scope.go:117] "RemoveContainer" containerID="f3b99ed4c156769f1e9dbd9952e637c4d0ffe19173eba5f1879fd8d46a3d92b6" Jan 21 18:11:07 crc kubenswrapper[4808]: I0121 18:11:07.458775 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vhqg2"] Jan 21 18:11:07 crc kubenswrapper[4808]: I0121 18:11:07.469353 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vhqg2"] Jan 21 18:11:07 crc kubenswrapper[4808]: I0121 18:11:07.483164 4808 scope.go:117] "RemoveContainer" containerID="21250bbf881927a048dad8a97c51c929f2ecb95507e06650ea92a84f9937aeb4" Jan 21 18:11:07 crc kubenswrapper[4808]: I0121 18:11:07.499112 4808 scope.go:117] "RemoveContainer" containerID="6abf2cdf457bf50e514b7a5c1646abd69c41b889cacaa4b6022179ffcc105e2c" Jan 21 18:11:07 crc kubenswrapper[4808]: E0121 18:11:07.499602 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6abf2cdf457bf50e514b7a5c1646abd69c41b889cacaa4b6022179ffcc105e2c\": container with ID starting with 6abf2cdf457bf50e514b7a5c1646abd69c41b889cacaa4b6022179ffcc105e2c not found: ID does not exist" containerID="6abf2cdf457bf50e514b7a5c1646abd69c41b889cacaa4b6022179ffcc105e2c" Jan 21 18:11:07 crc kubenswrapper[4808]: I0121 18:11:07.499632 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6abf2cdf457bf50e514b7a5c1646abd69c41b889cacaa4b6022179ffcc105e2c"} err="failed to get container status \"6abf2cdf457bf50e514b7a5c1646abd69c41b889cacaa4b6022179ffcc105e2c\": rpc error: code = NotFound desc = could not find container \"6abf2cdf457bf50e514b7a5c1646abd69c41b889cacaa4b6022179ffcc105e2c\": container with ID starting with 6abf2cdf457bf50e514b7a5c1646abd69c41b889cacaa4b6022179ffcc105e2c not found: ID does not exist" Jan 21 18:11:07 crc kubenswrapper[4808]: I0121 18:11:07.499653 4808 scope.go:117] "RemoveContainer" containerID="f3b99ed4c156769f1e9dbd9952e637c4d0ffe19173eba5f1879fd8d46a3d92b6" Jan 21 18:11:07 crc kubenswrapper[4808]: E0121 18:11:07.500084 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3b99ed4c156769f1e9dbd9952e637c4d0ffe19173eba5f1879fd8d46a3d92b6\": container with ID starting with f3b99ed4c156769f1e9dbd9952e637c4d0ffe19173eba5f1879fd8d46a3d92b6 not found: ID does not exist" containerID="f3b99ed4c156769f1e9dbd9952e637c4d0ffe19173eba5f1879fd8d46a3d92b6" Jan 21 18:11:07 crc kubenswrapper[4808]: I0121 18:11:07.500104 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3b99ed4c156769f1e9dbd9952e637c4d0ffe19173eba5f1879fd8d46a3d92b6"} err="failed to get container status \"f3b99ed4c156769f1e9dbd9952e637c4d0ffe19173eba5f1879fd8d46a3d92b6\": rpc error: code = NotFound desc = could not find container \"f3b99ed4c156769f1e9dbd9952e637c4d0ffe19173eba5f1879fd8d46a3d92b6\": container with ID starting with f3b99ed4c156769f1e9dbd9952e637c4d0ffe19173eba5f1879fd8d46a3d92b6 not found: ID does not exist" Jan 21 18:11:07 crc kubenswrapper[4808]: I0121 18:11:07.500118 4808 scope.go:117] "RemoveContainer" containerID="21250bbf881927a048dad8a97c51c929f2ecb95507e06650ea92a84f9937aeb4" Jan 21 18:11:07 crc kubenswrapper[4808]: E0121 18:11:07.500433 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21250bbf881927a048dad8a97c51c929f2ecb95507e06650ea92a84f9937aeb4\": container with ID starting with 21250bbf881927a048dad8a97c51c929f2ecb95507e06650ea92a84f9937aeb4 not found: ID does not exist" containerID="21250bbf881927a048dad8a97c51c929f2ecb95507e06650ea92a84f9937aeb4" Jan 21 18:11:07 crc kubenswrapper[4808]: I0121 18:11:07.500456 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21250bbf881927a048dad8a97c51c929f2ecb95507e06650ea92a84f9937aeb4"} err="failed to get container status \"21250bbf881927a048dad8a97c51c929f2ecb95507e06650ea92a84f9937aeb4\": rpc error: code = NotFound desc = could not find container \"21250bbf881927a048dad8a97c51c929f2ecb95507e06650ea92a84f9937aeb4\": container with ID starting with 21250bbf881927a048dad8a97c51c929f2ecb95507e06650ea92a84f9937aeb4 not found: ID does not exist" Jan 21 18:11:07 crc kubenswrapper[4808]: I0121 18:11:07.533416 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9667435d-cc17-4d75-8a83-3fbbd87ef20f" path="/var/lib/kubelet/pods/9667435d-cc17-4d75-8a83-3fbbd87ef20f/volumes" Jan 21 18:12:29 crc kubenswrapper[4808]: I0121 18:12:29.600307 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:12:29 crc kubenswrapper[4808]: I0121 18:12:29.600871 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:12:59 crc kubenswrapper[4808]: I0121 18:12:59.599756 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:12:59 crc kubenswrapper[4808]: I0121 18:12:59.601524 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:13:29 crc kubenswrapper[4808]: I0121 18:13:29.600168 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:13:29 crc kubenswrapper[4808]: I0121 18:13:29.600865 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:13:29 crc kubenswrapper[4808]: I0121 18:13:29.600917 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 18:13:29 crc kubenswrapper[4808]: I0121 18:13:29.601617 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"22ca6704d095a7830b6b1f1aa076b2fe0d7fab0a47c466111f510e13e0730cbe"} pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 18:13:29 crc kubenswrapper[4808]: I0121 18:13:29.601679 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" containerID="cri-o://22ca6704d095a7830b6b1f1aa076b2fe0d7fab0a47c466111f510e13e0730cbe" gracePeriod=600 Jan 21 18:13:30 crc kubenswrapper[4808]: I0121 18:13:30.444505 4808 generic.go:334] "Generic (PLEG): container finished" podID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerID="22ca6704d095a7830b6b1f1aa076b2fe0d7fab0a47c466111f510e13e0730cbe" exitCode=0 Jan 21 18:13:30 crc kubenswrapper[4808]: I0121 18:13:30.444576 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerDied","Data":"22ca6704d095a7830b6b1f1aa076b2fe0d7fab0a47c466111f510e13e0730cbe"} Jan 21 18:13:30 crc kubenswrapper[4808]: I0121 18:13:30.444873 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerStarted","Data":"c3613e51bff35c0c4f256683afea9976d826fc14c17f05a27e837d5940d33755"} Jan 21 18:13:30 crc kubenswrapper[4808]: I0121 18:13:30.444897 4808 scope.go:117] "RemoveContainer" containerID="5fd5f938b5c81db84b9cd3776775f788a436e3c24fc43c86adbeb98927f5200d" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.169938 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483655-lr2md"] Jan 21 18:15:00 crc kubenswrapper[4808]: E0121 18:15:00.171492 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25818535-6cfc-4cf2-abf5-be6e9408e9ad" containerName="registry-server" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.171515 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="25818535-6cfc-4cf2-abf5-be6e9408e9ad" containerName="registry-server" Jan 21 18:15:00 crc kubenswrapper[4808]: E0121 18:15:00.171537 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9affcfe2-bcd3-44ef-ad84-24978a483620" containerName="registry-server" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.171551 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9affcfe2-bcd3-44ef-ad84-24978a483620" containerName="registry-server" Jan 21 18:15:00 crc kubenswrapper[4808]: E0121 18:15:00.171567 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9667435d-cc17-4d75-8a83-3fbbd87ef20f" containerName="registry-server" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.171576 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9667435d-cc17-4d75-8a83-3fbbd87ef20f" containerName="registry-server" Jan 21 18:15:00 crc kubenswrapper[4808]: E0121 18:15:00.171591 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9affcfe2-bcd3-44ef-ad84-24978a483620" containerName="extract-content" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.171598 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9affcfe2-bcd3-44ef-ad84-24978a483620" containerName="extract-content" Jan 21 18:15:00 crc kubenswrapper[4808]: E0121 18:15:00.171612 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25818535-6cfc-4cf2-abf5-be6e9408e9ad" containerName="extract-content" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.171619 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="25818535-6cfc-4cf2-abf5-be6e9408e9ad" containerName="extract-content" Jan 21 18:15:00 crc kubenswrapper[4808]: E0121 18:15:00.171637 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9667435d-cc17-4d75-8a83-3fbbd87ef20f" containerName="extract-utilities" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.171646 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9667435d-cc17-4d75-8a83-3fbbd87ef20f" containerName="extract-utilities" Jan 21 18:15:00 crc kubenswrapper[4808]: E0121 18:15:00.171661 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9667435d-cc17-4d75-8a83-3fbbd87ef20f" containerName="extract-content" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.171671 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9667435d-cc17-4d75-8a83-3fbbd87ef20f" containerName="extract-content" Jan 21 18:15:00 crc kubenswrapper[4808]: E0121 18:15:00.171683 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25818535-6cfc-4cf2-abf5-be6e9408e9ad" containerName="extract-utilities" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.171692 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="25818535-6cfc-4cf2-abf5-be6e9408e9ad" containerName="extract-utilities" Jan 21 18:15:00 crc kubenswrapper[4808]: E0121 18:15:00.171706 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9affcfe2-bcd3-44ef-ad84-24978a483620" containerName="extract-utilities" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.171713 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9affcfe2-bcd3-44ef-ad84-24978a483620" containerName="extract-utilities" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.171897 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="25818535-6cfc-4cf2-abf5-be6e9408e9ad" containerName="registry-server" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.171909 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="9affcfe2-bcd3-44ef-ad84-24978a483620" containerName="registry-server" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.171919 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="9667435d-cc17-4d75-8a83-3fbbd87ef20f" containerName="registry-server" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.173932 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483655-lr2md" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.176300 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.178382 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.186566 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483655-lr2md"] Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.285502 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ckf6\" (UniqueName: \"kubernetes.io/projected/b5fe4500-1040-4b2d-926b-37dfb7fc03d0-kube-api-access-9ckf6\") pod \"collect-profiles-29483655-lr2md\" (UID: \"b5fe4500-1040-4b2d-926b-37dfb7fc03d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483655-lr2md" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.285663 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b5fe4500-1040-4b2d-926b-37dfb7fc03d0-secret-volume\") pod \"collect-profiles-29483655-lr2md\" (UID: \"b5fe4500-1040-4b2d-926b-37dfb7fc03d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483655-lr2md" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.285777 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b5fe4500-1040-4b2d-926b-37dfb7fc03d0-config-volume\") pod \"collect-profiles-29483655-lr2md\" (UID: \"b5fe4500-1040-4b2d-926b-37dfb7fc03d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483655-lr2md" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.387295 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b5fe4500-1040-4b2d-926b-37dfb7fc03d0-secret-volume\") pod \"collect-profiles-29483655-lr2md\" (UID: \"b5fe4500-1040-4b2d-926b-37dfb7fc03d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483655-lr2md" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.387350 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b5fe4500-1040-4b2d-926b-37dfb7fc03d0-config-volume\") pod \"collect-profiles-29483655-lr2md\" (UID: \"b5fe4500-1040-4b2d-926b-37dfb7fc03d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483655-lr2md" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.387375 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ckf6\" (UniqueName: \"kubernetes.io/projected/b5fe4500-1040-4b2d-926b-37dfb7fc03d0-kube-api-access-9ckf6\") pod \"collect-profiles-29483655-lr2md\" (UID: \"b5fe4500-1040-4b2d-926b-37dfb7fc03d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483655-lr2md" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.388863 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b5fe4500-1040-4b2d-926b-37dfb7fc03d0-config-volume\") pod \"collect-profiles-29483655-lr2md\" (UID: \"b5fe4500-1040-4b2d-926b-37dfb7fc03d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483655-lr2md" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.394210 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b5fe4500-1040-4b2d-926b-37dfb7fc03d0-secret-volume\") pod \"collect-profiles-29483655-lr2md\" (UID: \"b5fe4500-1040-4b2d-926b-37dfb7fc03d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483655-lr2md" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.403546 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ckf6\" (UniqueName: \"kubernetes.io/projected/b5fe4500-1040-4b2d-926b-37dfb7fc03d0-kube-api-access-9ckf6\") pod \"collect-profiles-29483655-lr2md\" (UID: \"b5fe4500-1040-4b2d-926b-37dfb7fc03d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483655-lr2md" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.493451 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483655-lr2md" Jan 21 18:15:00 crc kubenswrapper[4808]: I0121 18:15:00.697014 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483655-lr2md"] Jan 21 18:15:01 crc kubenswrapper[4808]: I0121 18:15:01.123014 4808 generic.go:334] "Generic (PLEG): container finished" podID="b5fe4500-1040-4b2d-926b-37dfb7fc03d0" containerID="e034467d84fdd7b128d57c97afd555c9d99db2df6f2611883b3c3de2da93e241" exitCode=0 Jan 21 18:15:01 crc kubenswrapper[4808]: I0121 18:15:01.123053 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483655-lr2md" event={"ID":"b5fe4500-1040-4b2d-926b-37dfb7fc03d0","Type":"ContainerDied","Data":"e034467d84fdd7b128d57c97afd555c9d99db2df6f2611883b3c3de2da93e241"} Jan 21 18:15:01 crc kubenswrapper[4808]: I0121 18:15:01.123082 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483655-lr2md" event={"ID":"b5fe4500-1040-4b2d-926b-37dfb7fc03d0","Type":"ContainerStarted","Data":"21967e0ef713ce5d3917db2759232e742683489484f581cb4601d9cf140d1a1a"} Jan 21 18:15:02 crc kubenswrapper[4808]: I0121 18:15:02.371649 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483655-lr2md" Jan 21 18:15:02 crc kubenswrapper[4808]: I0121 18:15:02.427739 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b5fe4500-1040-4b2d-926b-37dfb7fc03d0-config-volume\") pod \"b5fe4500-1040-4b2d-926b-37dfb7fc03d0\" (UID: \"b5fe4500-1040-4b2d-926b-37dfb7fc03d0\") " Jan 21 18:15:02 crc kubenswrapper[4808]: I0121 18:15:02.427779 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b5fe4500-1040-4b2d-926b-37dfb7fc03d0-secret-volume\") pod \"b5fe4500-1040-4b2d-926b-37dfb7fc03d0\" (UID: \"b5fe4500-1040-4b2d-926b-37dfb7fc03d0\") " Jan 21 18:15:02 crc kubenswrapper[4808]: I0121 18:15:02.427843 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ckf6\" (UniqueName: \"kubernetes.io/projected/b5fe4500-1040-4b2d-926b-37dfb7fc03d0-kube-api-access-9ckf6\") pod \"b5fe4500-1040-4b2d-926b-37dfb7fc03d0\" (UID: \"b5fe4500-1040-4b2d-926b-37dfb7fc03d0\") " Jan 21 18:15:02 crc kubenswrapper[4808]: I0121 18:15:02.428593 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5fe4500-1040-4b2d-926b-37dfb7fc03d0-config-volume" (OuterVolumeSpecName: "config-volume") pod "b5fe4500-1040-4b2d-926b-37dfb7fc03d0" (UID: "b5fe4500-1040-4b2d-926b-37dfb7fc03d0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:15:02 crc kubenswrapper[4808]: I0121 18:15:02.432589 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5fe4500-1040-4b2d-926b-37dfb7fc03d0-kube-api-access-9ckf6" (OuterVolumeSpecName: "kube-api-access-9ckf6") pod "b5fe4500-1040-4b2d-926b-37dfb7fc03d0" (UID: "b5fe4500-1040-4b2d-926b-37dfb7fc03d0"). InnerVolumeSpecName "kube-api-access-9ckf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:15:02 crc kubenswrapper[4808]: I0121 18:15:02.432589 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5fe4500-1040-4b2d-926b-37dfb7fc03d0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b5fe4500-1040-4b2d-926b-37dfb7fc03d0" (UID: "b5fe4500-1040-4b2d-926b-37dfb7fc03d0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:15:02 crc kubenswrapper[4808]: I0121 18:15:02.530227 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ckf6\" (UniqueName: \"kubernetes.io/projected/b5fe4500-1040-4b2d-926b-37dfb7fc03d0-kube-api-access-9ckf6\") on node \"crc\" DevicePath \"\"" Jan 21 18:15:02 crc kubenswrapper[4808]: I0121 18:15:02.530307 4808 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b5fe4500-1040-4b2d-926b-37dfb7fc03d0-config-volume\") on node \"crc\" DevicePath \"\"" Jan 21 18:15:02 crc kubenswrapper[4808]: I0121 18:15:02.530327 4808 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b5fe4500-1040-4b2d-926b-37dfb7fc03d0-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 21 18:15:03 crc kubenswrapper[4808]: I0121 18:15:03.137609 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483655-lr2md" Jan 21 18:15:03 crc kubenswrapper[4808]: I0121 18:15:03.137605 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483655-lr2md" event={"ID":"b5fe4500-1040-4b2d-926b-37dfb7fc03d0","Type":"ContainerDied","Data":"21967e0ef713ce5d3917db2759232e742683489484f581cb4601d9cf140d1a1a"} Jan 21 18:15:03 crc kubenswrapper[4808]: I0121 18:15:03.137770 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21967e0ef713ce5d3917db2759232e742683489484f581cb4601d9cf140d1a1a" Jan 21 18:15:29 crc kubenswrapper[4808]: I0121 18:15:29.599920 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:15:29 crc kubenswrapper[4808]: I0121 18:15:29.600446 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:15:59 crc kubenswrapper[4808]: I0121 18:15:59.600296 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:15:59 crc kubenswrapper[4808]: I0121 18:15:59.601021 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:16:29 crc kubenswrapper[4808]: I0121 18:16:29.599860 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:16:29 crc kubenswrapper[4808]: I0121 18:16:29.601818 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:16:29 crc kubenswrapper[4808]: I0121 18:16:29.601932 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 18:16:29 crc kubenswrapper[4808]: I0121 18:16:29.602591 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c3613e51bff35c0c4f256683afea9976d826fc14c17f05a27e837d5940d33755"} pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 18:16:29 crc kubenswrapper[4808]: I0121 18:16:29.602731 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" containerID="cri-o://c3613e51bff35c0c4f256683afea9976d826fc14c17f05a27e837d5940d33755" gracePeriod=600 Jan 21 18:16:29 crc kubenswrapper[4808]: I0121 18:16:29.793602 4808 generic.go:334] "Generic (PLEG): container finished" podID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerID="c3613e51bff35c0c4f256683afea9976d826fc14c17f05a27e837d5940d33755" exitCode=0 Jan 21 18:16:29 crc kubenswrapper[4808]: I0121 18:16:29.793657 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerDied","Data":"c3613e51bff35c0c4f256683afea9976d826fc14c17f05a27e837d5940d33755"} Jan 21 18:16:29 crc kubenswrapper[4808]: I0121 18:16:29.793972 4808 scope.go:117] "RemoveContainer" containerID="22ca6704d095a7830b6b1f1aa076b2fe0d7fab0a47c466111f510e13e0730cbe" Jan 21 18:16:31 crc kubenswrapper[4808]: I0121 18:16:31.811355 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerStarted","Data":"fdbd789ae8672bd0f6357c9d20a4c92b6be5b8f6f3df753f537861c9d45d3858"} Jan 21 18:18:59 crc kubenswrapper[4808]: I0121 18:18:59.599737 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:18:59 crc kubenswrapper[4808]: I0121 18:18:59.600412 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:19:29 crc kubenswrapper[4808]: I0121 18:19:29.599529 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:19:29 crc kubenswrapper[4808]: I0121 18:19:29.600067 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:19:59 crc kubenswrapper[4808]: I0121 18:19:59.599693 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:19:59 crc kubenswrapper[4808]: I0121 18:19:59.600705 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:19:59 crc kubenswrapper[4808]: I0121 18:19:59.600786 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 18:19:59 crc kubenswrapper[4808]: I0121 18:19:59.601696 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fdbd789ae8672bd0f6357c9d20a4c92b6be5b8f6f3df753f537861c9d45d3858"} pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 18:19:59 crc kubenswrapper[4808]: I0121 18:19:59.601780 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" containerID="cri-o://fdbd789ae8672bd0f6357c9d20a4c92b6be5b8f6f3df753f537861c9d45d3858" gracePeriod=600 Jan 21 18:20:00 crc kubenswrapper[4808]: I0121 18:20:00.278853 4808 generic.go:334] "Generic (PLEG): container finished" podID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerID="fdbd789ae8672bd0f6357c9d20a4c92b6be5b8f6f3df753f537861c9d45d3858" exitCode=0 Jan 21 18:20:00 crc kubenswrapper[4808]: I0121 18:20:00.278927 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerDied","Data":"fdbd789ae8672bd0f6357c9d20a4c92b6be5b8f6f3df753f537861c9d45d3858"} Jan 21 18:20:00 crc kubenswrapper[4808]: I0121 18:20:00.279257 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerStarted","Data":"3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1"} Jan 21 18:20:00 crc kubenswrapper[4808]: I0121 18:20:00.279285 4808 scope.go:117] "RemoveContainer" containerID="c3613e51bff35c0c4f256683afea9976d826fc14c17f05a27e837d5940d33755" Jan 21 18:21:04 crc kubenswrapper[4808]: I0121 18:21:04.972085 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dnttl"] Jan 21 18:21:04 crc kubenswrapper[4808]: E0121 18:21:04.973517 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5fe4500-1040-4b2d-926b-37dfb7fc03d0" containerName="collect-profiles" Jan 21 18:21:04 crc kubenswrapper[4808]: I0121 18:21:04.973537 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5fe4500-1040-4b2d-926b-37dfb7fc03d0" containerName="collect-profiles" Jan 21 18:21:04 crc kubenswrapper[4808]: I0121 18:21:04.973680 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5fe4500-1040-4b2d-926b-37dfb7fc03d0" containerName="collect-profiles" Jan 21 18:21:04 crc kubenswrapper[4808]: I0121 18:21:04.975448 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dnttl" Jan 21 18:21:04 crc kubenswrapper[4808]: I0121 18:21:04.979722 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/928a038e-621a-4f7c-822e-cc3b50799fe5-catalog-content\") pod \"certified-operators-dnttl\" (UID: \"928a038e-621a-4f7c-822e-cc3b50799fe5\") " pod="openshift-marketplace/certified-operators-dnttl" Jan 21 18:21:04 crc kubenswrapper[4808]: I0121 18:21:04.979809 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/928a038e-621a-4f7c-822e-cc3b50799fe5-utilities\") pod \"certified-operators-dnttl\" (UID: \"928a038e-621a-4f7c-822e-cc3b50799fe5\") " pod="openshift-marketplace/certified-operators-dnttl" Jan 21 18:21:04 crc kubenswrapper[4808]: I0121 18:21:04.979844 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q97hb\" (UniqueName: \"kubernetes.io/projected/928a038e-621a-4f7c-822e-cc3b50799fe5-kube-api-access-q97hb\") pod \"certified-operators-dnttl\" (UID: \"928a038e-621a-4f7c-822e-cc3b50799fe5\") " pod="openshift-marketplace/certified-operators-dnttl" Jan 21 18:21:04 crc kubenswrapper[4808]: I0121 18:21:04.980096 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dnttl"] Jan 21 18:21:05 crc kubenswrapper[4808]: I0121 18:21:05.081335 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/928a038e-621a-4f7c-822e-cc3b50799fe5-utilities\") pod \"certified-operators-dnttl\" (UID: \"928a038e-621a-4f7c-822e-cc3b50799fe5\") " pod="openshift-marketplace/certified-operators-dnttl" Jan 21 18:21:05 crc kubenswrapper[4808]: I0121 18:21:05.081391 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q97hb\" (UniqueName: \"kubernetes.io/projected/928a038e-621a-4f7c-822e-cc3b50799fe5-kube-api-access-q97hb\") pod \"certified-operators-dnttl\" (UID: \"928a038e-621a-4f7c-822e-cc3b50799fe5\") " pod="openshift-marketplace/certified-operators-dnttl" Jan 21 18:21:05 crc kubenswrapper[4808]: I0121 18:21:05.081494 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/928a038e-621a-4f7c-822e-cc3b50799fe5-catalog-content\") pod \"certified-operators-dnttl\" (UID: \"928a038e-621a-4f7c-822e-cc3b50799fe5\") " pod="openshift-marketplace/certified-operators-dnttl" Jan 21 18:21:05 crc kubenswrapper[4808]: I0121 18:21:05.081821 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/928a038e-621a-4f7c-822e-cc3b50799fe5-utilities\") pod \"certified-operators-dnttl\" (UID: \"928a038e-621a-4f7c-822e-cc3b50799fe5\") " pod="openshift-marketplace/certified-operators-dnttl" Jan 21 18:21:05 crc kubenswrapper[4808]: I0121 18:21:05.081880 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/928a038e-621a-4f7c-822e-cc3b50799fe5-catalog-content\") pod \"certified-operators-dnttl\" (UID: \"928a038e-621a-4f7c-822e-cc3b50799fe5\") " pod="openshift-marketplace/certified-operators-dnttl" Jan 21 18:21:05 crc kubenswrapper[4808]: I0121 18:21:05.110971 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q97hb\" (UniqueName: \"kubernetes.io/projected/928a038e-621a-4f7c-822e-cc3b50799fe5-kube-api-access-q97hb\") pod \"certified-operators-dnttl\" (UID: \"928a038e-621a-4f7c-822e-cc3b50799fe5\") " pod="openshift-marketplace/certified-operators-dnttl" Jan 21 18:21:05 crc kubenswrapper[4808]: I0121 18:21:05.305869 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dnttl" Jan 21 18:21:05 crc kubenswrapper[4808]: I0121 18:21:05.562914 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dnttl"] Jan 21 18:21:05 crc kubenswrapper[4808]: I0121 18:21:05.738933 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnttl" event={"ID":"928a038e-621a-4f7c-822e-cc3b50799fe5","Type":"ContainerStarted","Data":"9110c966dac86e7478dda447c39192bf59a464a338ce3a64ea3af50cea41eae7"} Jan 21 18:21:06 crc kubenswrapper[4808]: I0121 18:21:06.746546 4808 generic.go:334] "Generic (PLEG): container finished" podID="928a038e-621a-4f7c-822e-cc3b50799fe5" containerID="e7385dfa3efd82c8af9ab3d27cc455dd1c273ae88ac78c2df5b8a646293f86b9" exitCode=0 Jan 21 18:21:06 crc kubenswrapper[4808]: I0121 18:21:06.746623 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnttl" event={"ID":"928a038e-621a-4f7c-822e-cc3b50799fe5","Type":"ContainerDied","Data":"e7385dfa3efd82c8af9ab3d27cc455dd1c273ae88ac78c2df5b8a646293f86b9"} Jan 21 18:21:06 crc kubenswrapper[4808]: I0121 18:21:06.752344 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 18:21:07 crc kubenswrapper[4808]: I0121 18:21:07.754689 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnttl" event={"ID":"928a038e-621a-4f7c-822e-cc3b50799fe5","Type":"ContainerStarted","Data":"e858a65136f3fa1564f2b9a2468e725cef1942f66c529e73572be8624f40cffe"} Jan 21 18:21:08 crc kubenswrapper[4808]: I0121 18:21:08.564136 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4fgg7"] Jan 21 18:21:08 crc kubenswrapper[4808]: I0121 18:21:08.566124 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4fgg7" Jan 21 18:21:08 crc kubenswrapper[4808]: I0121 18:21:08.578516 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4fgg7"] Jan 21 18:21:08 crc kubenswrapper[4808]: I0121 18:21:08.740076 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e185876e-8a0f-4583-ab42-6325279d31d2-utilities\") pod \"community-operators-4fgg7\" (UID: \"e185876e-8a0f-4583-ab42-6325279d31d2\") " pod="openshift-marketplace/community-operators-4fgg7" Jan 21 18:21:08 crc kubenswrapper[4808]: I0121 18:21:08.740189 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e185876e-8a0f-4583-ab42-6325279d31d2-catalog-content\") pod \"community-operators-4fgg7\" (UID: \"e185876e-8a0f-4583-ab42-6325279d31d2\") " pod="openshift-marketplace/community-operators-4fgg7" Jan 21 18:21:08 crc kubenswrapper[4808]: I0121 18:21:08.740219 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zgfx\" (UniqueName: \"kubernetes.io/projected/e185876e-8a0f-4583-ab42-6325279d31d2-kube-api-access-9zgfx\") pod \"community-operators-4fgg7\" (UID: \"e185876e-8a0f-4583-ab42-6325279d31d2\") " pod="openshift-marketplace/community-operators-4fgg7" Jan 21 18:21:08 crc kubenswrapper[4808]: I0121 18:21:08.765061 4808 generic.go:334] "Generic (PLEG): container finished" podID="928a038e-621a-4f7c-822e-cc3b50799fe5" containerID="e858a65136f3fa1564f2b9a2468e725cef1942f66c529e73572be8624f40cffe" exitCode=0 Jan 21 18:21:08 crc kubenswrapper[4808]: I0121 18:21:08.765100 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnttl" event={"ID":"928a038e-621a-4f7c-822e-cc3b50799fe5","Type":"ContainerDied","Data":"e858a65136f3fa1564f2b9a2468e725cef1942f66c529e73572be8624f40cffe"} Jan 21 18:21:08 crc kubenswrapper[4808]: I0121 18:21:08.841625 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zgfx\" (UniqueName: \"kubernetes.io/projected/e185876e-8a0f-4583-ab42-6325279d31d2-kube-api-access-9zgfx\") pod \"community-operators-4fgg7\" (UID: \"e185876e-8a0f-4583-ab42-6325279d31d2\") " pod="openshift-marketplace/community-operators-4fgg7" Jan 21 18:21:08 crc kubenswrapper[4808]: I0121 18:21:08.842091 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e185876e-8a0f-4583-ab42-6325279d31d2-utilities\") pod \"community-operators-4fgg7\" (UID: \"e185876e-8a0f-4583-ab42-6325279d31d2\") " pod="openshift-marketplace/community-operators-4fgg7" Jan 21 18:21:08 crc kubenswrapper[4808]: I0121 18:21:08.842317 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e185876e-8a0f-4583-ab42-6325279d31d2-catalog-content\") pod \"community-operators-4fgg7\" (UID: \"e185876e-8a0f-4583-ab42-6325279d31d2\") " pod="openshift-marketplace/community-operators-4fgg7" Jan 21 18:21:08 crc kubenswrapper[4808]: I0121 18:21:08.842816 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e185876e-8a0f-4583-ab42-6325279d31d2-catalog-content\") pod \"community-operators-4fgg7\" (UID: \"e185876e-8a0f-4583-ab42-6325279d31d2\") " pod="openshift-marketplace/community-operators-4fgg7" Jan 21 18:21:08 crc kubenswrapper[4808]: I0121 18:21:08.843093 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e185876e-8a0f-4583-ab42-6325279d31d2-utilities\") pod \"community-operators-4fgg7\" (UID: \"e185876e-8a0f-4583-ab42-6325279d31d2\") " pod="openshift-marketplace/community-operators-4fgg7" Jan 21 18:21:08 crc kubenswrapper[4808]: I0121 18:21:08.878677 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zgfx\" (UniqueName: \"kubernetes.io/projected/e185876e-8a0f-4583-ab42-6325279d31d2-kube-api-access-9zgfx\") pod \"community-operators-4fgg7\" (UID: \"e185876e-8a0f-4583-ab42-6325279d31d2\") " pod="openshift-marketplace/community-operators-4fgg7" Jan 21 18:21:08 crc kubenswrapper[4808]: I0121 18:21:08.888035 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4fgg7" Jan 21 18:21:09 crc kubenswrapper[4808]: I0121 18:21:09.205064 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4fgg7"] Jan 21 18:21:09 crc kubenswrapper[4808]: W0121 18:21:09.211426 4808 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode185876e_8a0f_4583_ab42_6325279d31d2.slice/crio-b3eb3cb56d53b58f5d989344a0aa416ec7a44908b739a1c2c5a08469ea79e2cd WatchSource:0}: Error finding container b3eb3cb56d53b58f5d989344a0aa416ec7a44908b739a1c2c5a08469ea79e2cd: Status 404 returned error can't find the container with id b3eb3cb56d53b58f5d989344a0aa416ec7a44908b739a1c2c5a08469ea79e2cd Jan 21 18:21:09 crc kubenswrapper[4808]: I0121 18:21:09.773456 4808 generic.go:334] "Generic (PLEG): container finished" podID="e185876e-8a0f-4583-ab42-6325279d31d2" containerID="c9cf3b9cbe6a1237275ad73ff52b8c00835a082571aa04ceea4f0dfea0f5e631" exitCode=0 Jan 21 18:21:09 crc kubenswrapper[4808]: I0121 18:21:09.773570 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4fgg7" event={"ID":"e185876e-8a0f-4583-ab42-6325279d31d2","Type":"ContainerDied","Data":"c9cf3b9cbe6a1237275ad73ff52b8c00835a082571aa04ceea4f0dfea0f5e631"} Jan 21 18:21:09 crc kubenswrapper[4808]: I0121 18:21:09.774118 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4fgg7" event={"ID":"e185876e-8a0f-4583-ab42-6325279d31d2","Type":"ContainerStarted","Data":"b3eb3cb56d53b58f5d989344a0aa416ec7a44908b739a1c2c5a08469ea79e2cd"} Jan 21 18:21:09 crc kubenswrapper[4808]: I0121 18:21:09.780112 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnttl" event={"ID":"928a038e-621a-4f7c-822e-cc3b50799fe5","Type":"ContainerStarted","Data":"998cb4a6772b6f372e00582b3fc58a080a488732fcba621c7e16f1c0c12c3595"} Jan 21 18:21:09 crc kubenswrapper[4808]: I0121 18:21:09.814520 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dnttl" podStartSLOduration=3.159601148 podStartE2EDuration="5.814501688s" podCreationTimestamp="2026-01-21 18:21:04 +0000 UTC" firstStartedPulling="2026-01-21 18:21:06.752008101 +0000 UTC m=+1501.482518986" lastFinishedPulling="2026-01-21 18:21:09.406908641 +0000 UTC m=+1504.137419526" observedRunningTime="2026-01-21 18:21:09.807119602 +0000 UTC m=+1504.537630487" watchObservedRunningTime="2026-01-21 18:21:09.814501688 +0000 UTC m=+1504.545012563" Jan 21 18:21:10 crc kubenswrapper[4808]: I0121 18:21:10.787182 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4fgg7" event={"ID":"e185876e-8a0f-4583-ab42-6325279d31d2","Type":"ContainerStarted","Data":"3cc2537dcb2c2e48bf9b07854c90e9358a1937ae6615ed7dbc1e329af32fdabf"} Jan 21 18:21:11 crc kubenswrapper[4808]: I0121 18:21:11.796456 4808 generic.go:334] "Generic (PLEG): container finished" podID="e185876e-8a0f-4583-ab42-6325279d31d2" containerID="3cc2537dcb2c2e48bf9b07854c90e9358a1937ae6615ed7dbc1e329af32fdabf" exitCode=0 Jan 21 18:21:11 crc kubenswrapper[4808]: I0121 18:21:11.796537 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4fgg7" event={"ID":"e185876e-8a0f-4583-ab42-6325279d31d2","Type":"ContainerDied","Data":"3cc2537dcb2c2e48bf9b07854c90e9358a1937ae6615ed7dbc1e329af32fdabf"} Jan 21 18:21:12 crc kubenswrapper[4808]: I0121 18:21:12.805330 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4fgg7" event={"ID":"e185876e-8a0f-4583-ab42-6325279d31d2","Type":"ContainerStarted","Data":"c18c7550d4df91c72f900b07a8b4caf140151c403034c48c786c3dbc3ff0ddb6"} Jan 21 18:21:12 crc kubenswrapper[4808]: I0121 18:21:12.831065 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4fgg7" podStartSLOduration=2.333749818 podStartE2EDuration="4.831046414s" podCreationTimestamp="2026-01-21 18:21:08 +0000 UTC" firstStartedPulling="2026-01-21 18:21:09.775325368 +0000 UTC m=+1504.505836253" lastFinishedPulling="2026-01-21 18:21:12.272621964 +0000 UTC m=+1507.003132849" observedRunningTime="2026-01-21 18:21:12.827955531 +0000 UTC m=+1507.558466416" watchObservedRunningTime="2026-01-21 18:21:12.831046414 +0000 UTC m=+1507.561557289" Jan 21 18:21:15 crc kubenswrapper[4808]: I0121 18:21:15.306980 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dnttl" Jan 21 18:21:15 crc kubenswrapper[4808]: I0121 18:21:15.307908 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dnttl" Jan 21 18:21:15 crc kubenswrapper[4808]: I0121 18:21:15.350746 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dnttl" Jan 21 18:21:15 crc kubenswrapper[4808]: I0121 18:21:15.864834 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dnttl" Jan 21 18:21:16 crc kubenswrapper[4808]: I0121 18:21:16.543182 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dnttl"] Jan 21 18:21:17 crc kubenswrapper[4808]: I0121 18:21:17.835784 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dnttl" podUID="928a038e-621a-4f7c-822e-cc3b50799fe5" containerName="registry-server" containerID="cri-o://998cb4a6772b6f372e00582b3fc58a080a488732fcba621c7e16f1c0c12c3595" gracePeriod=2 Jan 21 18:21:18 crc kubenswrapper[4808]: I0121 18:21:18.889361 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4fgg7" Jan 21 18:21:18 crc kubenswrapper[4808]: I0121 18:21:18.889669 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4fgg7" Jan 21 18:21:18 crc kubenswrapper[4808]: I0121 18:21:18.932030 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4fgg7" Jan 21 18:21:19 crc kubenswrapper[4808]: I0121 18:21:19.850270 4808 generic.go:334] "Generic (PLEG): container finished" podID="928a038e-621a-4f7c-822e-cc3b50799fe5" containerID="998cb4a6772b6f372e00582b3fc58a080a488732fcba621c7e16f1c0c12c3595" exitCode=0 Jan 21 18:21:19 crc kubenswrapper[4808]: I0121 18:21:19.851847 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnttl" event={"ID":"928a038e-621a-4f7c-822e-cc3b50799fe5","Type":"ContainerDied","Data":"998cb4a6772b6f372e00582b3fc58a080a488732fcba621c7e16f1c0c12c3595"} Jan 21 18:21:19 crc kubenswrapper[4808]: I0121 18:21:19.906984 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4fgg7" Jan 21 18:21:19 crc kubenswrapper[4808]: I0121 18:21:19.960450 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4fgg7"] Jan 21 18:21:20 crc kubenswrapper[4808]: I0121 18:21:20.043311 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dnttl" Jan 21 18:21:20 crc kubenswrapper[4808]: I0121 18:21:20.156503 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q97hb\" (UniqueName: \"kubernetes.io/projected/928a038e-621a-4f7c-822e-cc3b50799fe5-kube-api-access-q97hb\") pod \"928a038e-621a-4f7c-822e-cc3b50799fe5\" (UID: \"928a038e-621a-4f7c-822e-cc3b50799fe5\") " Jan 21 18:21:20 crc kubenswrapper[4808]: I0121 18:21:20.156570 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/928a038e-621a-4f7c-822e-cc3b50799fe5-utilities\") pod \"928a038e-621a-4f7c-822e-cc3b50799fe5\" (UID: \"928a038e-621a-4f7c-822e-cc3b50799fe5\") " Jan 21 18:21:20 crc kubenswrapper[4808]: I0121 18:21:20.156604 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/928a038e-621a-4f7c-822e-cc3b50799fe5-catalog-content\") pod \"928a038e-621a-4f7c-822e-cc3b50799fe5\" (UID: \"928a038e-621a-4f7c-822e-cc3b50799fe5\") " Jan 21 18:21:20 crc kubenswrapper[4808]: I0121 18:21:20.157595 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/928a038e-621a-4f7c-822e-cc3b50799fe5-utilities" (OuterVolumeSpecName: "utilities") pod "928a038e-621a-4f7c-822e-cc3b50799fe5" (UID: "928a038e-621a-4f7c-822e-cc3b50799fe5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:21:20 crc kubenswrapper[4808]: I0121 18:21:20.180441 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/928a038e-621a-4f7c-822e-cc3b50799fe5-kube-api-access-q97hb" (OuterVolumeSpecName: "kube-api-access-q97hb") pod "928a038e-621a-4f7c-822e-cc3b50799fe5" (UID: "928a038e-621a-4f7c-822e-cc3b50799fe5"). InnerVolumeSpecName "kube-api-access-q97hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:21:20 crc kubenswrapper[4808]: I0121 18:21:20.204515 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/928a038e-621a-4f7c-822e-cc3b50799fe5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "928a038e-621a-4f7c-822e-cc3b50799fe5" (UID: "928a038e-621a-4f7c-822e-cc3b50799fe5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:21:20 crc kubenswrapper[4808]: I0121 18:21:20.258406 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q97hb\" (UniqueName: \"kubernetes.io/projected/928a038e-621a-4f7c-822e-cc3b50799fe5-kube-api-access-q97hb\") on node \"crc\" DevicePath \"\"" Jan 21 18:21:20 crc kubenswrapper[4808]: I0121 18:21:20.258439 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/928a038e-621a-4f7c-822e-cc3b50799fe5-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:21:20 crc kubenswrapper[4808]: I0121 18:21:20.258449 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/928a038e-621a-4f7c-822e-cc3b50799fe5-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:21:20 crc kubenswrapper[4808]: I0121 18:21:20.857876 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dnttl" Jan 21 18:21:20 crc kubenswrapper[4808]: I0121 18:21:20.864359 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnttl" event={"ID":"928a038e-621a-4f7c-822e-cc3b50799fe5","Type":"ContainerDied","Data":"9110c966dac86e7478dda447c39192bf59a464a338ce3a64ea3af50cea41eae7"} Jan 21 18:21:20 crc kubenswrapper[4808]: I0121 18:21:20.864448 4808 scope.go:117] "RemoveContainer" containerID="998cb4a6772b6f372e00582b3fc58a080a488732fcba621c7e16f1c0c12c3595" Jan 21 18:21:20 crc kubenswrapper[4808]: I0121 18:21:20.886152 4808 scope.go:117] "RemoveContainer" containerID="e858a65136f3fa1564f2b9a2468e725cef1942f66c529e73572be8624f40cffe" Jan 21 18:21:20 crc kubenswrapper[4808]: I0121 18:21:20.891171 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dnttl"] Jan 21 18:21:20 crc kubenswrapper[4808]: I0121 18:21:20.896683 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dnttl"] Jan 21 18:21:20 crc kubenswrapper[4808]: I0121 18:21:20.927646 4808 scope.go:117] "RemoveContainer" containerID="e7385dfa3efd82c8af9ab3d27cc455dd1c273ae88ac78c2df5b8a646293f86b9" Jan 21 18:21:21 crc kubenswrapper[4808]: I0121 18:21:21.533181 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="928a038e-621a-4f7c-822e-cc3b50799fe5" path="/var/lib/kubelet/pods/928a038e-621a-4f7c-822e-cc3b50799fe5/volumes" Jan 21 18:21:21 crc kubenswrapper[4808]: I0121 18:21:21.865533 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4fgg7" podUID="e185876e-8a0f-4583-ab42-6325279d31d2" containerName="registry-server" containerID="cri-o://c18c7550d4df91c72f900b07a8b4caf140151c403034c48c786c3dbc3ff0ddb6" gracePeriod=2 Jan 21 18:21:23 crc kubenswrapper[4808]: I0121 18:21:23.885010 4808 generic.go:334] "Generic (PLEG): container finished" podID="e185876e-8a0f-4583-ab42-6325279d31d2" containerID="c18c7550d4df91c72f900b07a8b4caf140151c403034c48c786c3dbc3ff0ddb6" exitCode=0 Jan 21 18:21:23 crc kubenswrapper[4808]: I0121 18:21:23.885299 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4fgg7" event={"ID":"e185876e-8a0f-4583-ab42-6325279d31d2","Type":"ContainerDied","Data":"c18c7550d4df91c72f900b07a8b4caf140151c403034c48c786c3dbc3ff0ddb6"} Jan 21 18:21:24 crc kubenswrapper[4808]: I0121 18:21:24.030578 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4fgg7" Jan 21 18:21:24 crc kubenswrapper[4808]: I0121 18:21:24.112122 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e185876e-8a0f-4583-ab42-6325279d31d2-utilities\") pod \"e185876e-8a0f-4583-ab42-6325279d31d2\" (UID: \"e185876e-8a0f-4583-ab42-6325279d31d2\") " Jan 21 18:21:24 crc kubenswrapper[4808]: I0121 18:21:24.112159 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e185876e-8a0f-4583-ab42-6325279d31d2-catalog-content\") pod \"e185876e-8a0f-4583-ab42-6325279d31d2\" (UID: \"e185876e-8a0f-4583-ab42-6325279d31d2\") " Jan 21 18:21:24 crc kubenswrapper[4808]: I0121 18:21:24.112204 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zgfx\" (UniqueName: \"kubernetes.io/projected/e185876e-8a0f-4583-ab42-6325279d31d2-kube-api-access-9zgfx\") pod \"e185876e-8a0f-4583-ab42-6325279d31d2\" (UID: \"e185876e-8a0f-4583-ab42-6325279d31d2\") " Jan 21 18:21:24 crc kubenswrapper[4808]: I0121 18:21:24.113036 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e185876e-8a0f-4583-ab42-6325279d31d2-utilities" (OuterVolumeSpecName: "utilities") pod "e185876e-8a0f-4583-ab42-6325279d31d2" (UID: "e185876e-8a0f-4583-ab42-6325279d31d2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:21:24 crc kubenswrapper[4808]: I0121 18:21:24.117281 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e185876e-8a0f-4583-ab42-6325279d31d2-kube-api-access-9zgfx" (OuterVolumeSpecName: "kube-api-access-9zgfx") pod "e185876e-8a0f-4583-ab42-6325279d31d2" (UID: "e185876e-8a0f-4583-ab42-6325279d31d2"). InnerVolumeSpecName "kube-api-access-9zgfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:21:24 crc kubenswrapper[4808]: I0121 18:21:24.172406 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e185876e-8a0f-4583-ab42-6325279d31d2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e185876e-8a0f-4583-ab42-6325279d31d2" (UID: "e185876e-8a0f-4583-ab42-6325279d31d2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:21:24 crc kubenswrapper[4808]: I0121 18:21:24.214028 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e185876e-8a0f-4583-ab42-6325279d31d2-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:21:24 crc kubenswrapper[4808]: I0121 18:21:24.214063 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e185876e-8a0f-4583-ab42-6325279d31d2-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:21:24 crc kubenswrapper[4808]: I0121 18:21:24.214073 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zgfx\" (UniqueName: \"kubernetes.io/projected/e185876e-8a0f-4583-ab42-6325279d31d2-kube-api-access-9zgfx\") on node \"crc\" DevicePath \"\"" Jan 21 18:21:24 crc kubenswrapper[4808]: I0121 18:21:24.894622 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4fgg7" event={"ID":"e185876e-8a0f-4583-ab42-6325279d31d2","Type":"ContainerDied","Data":"b3eb3cb56d53b58f5d989344a0aa416ec7a44908b739a1c2c5a08469ea79e2cd"} Jan 21 18:21:24 crc kubenswrapper[4808]: I0121 18:21:24.894669 4808 scope.go:117] "RemoveContainer" containerID="c18c7550d4df91c72f900b07a8b4caf140151c403034c48c786c3dbc3ff0ddb6" Jan 21 18:21:24 crc kubenswrapper[4808]: I0121 18:21:24.894713 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4fgg7" Jan 21 18:21:24 crc kubenswrapper[4808]: I0121 18:21:24.914955 4808 scope.go:117] "RemoveContainer" containerID="3cc2537dcb2c2e48bf9b07854c90e9358a1937ae6615ed7dbc1e329af32fdabf" Jan 21 18:21:24 crc kubenswrapper[4808]: I0121 18:21:24.923163 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4fgg7"] Jan 21 18:21:24 crc kubenswrapper[4808]: I0121 18:21:24.930432 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4fgg7"] Jan 21 18:21:24 crc kubenswrapper[4808]: I0121 18:21:24.941310 4808 scope.go:117] "RemoveContainer" containerID="c9cf3b9cbe6a1237275ad73ff52b8c00835a082571aa04ceea4f0dfea0f5e631" Jan 21 18:21:25 crc kubenswrapper[4808]: I0121 18:21:25.530205 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e185876e-8a0f-4583-ab42-6325279d31d2" path="/var/lib/kubelet/pods/e185876e-8a0f-4583-ab42-6325279d31d2/volumes" Jan 21 18:21:59 crc kubenswrapper[4808]: I0121 18:21:59.600168 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:21:59 crc kubenswrapper[4808]: I0121 18:21:59.600776 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:22:29 crc kubenswrapper[4808]: I0121 18:22:29.599608 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:22:29 crc kubenswrapper[4808]: I0121 18:22:29.601747 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:22:59 crc kubenswrapper[4808]: I0121 18:22:59.599330 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:22:59 crc kubenswrapper[4808]: I0121 18:22:59.599874 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:22:59 crc kubenswrapper[4808]: I0121 18:22:59.599915 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 18:22:59 crc kubenswrapper[4808]: I0121 18:22:59.600469 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1"} pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 18:22:59 crc kubenswrapper[4808]: I0121 18:22:59.600522 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" containerID="cri-o://3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" gracePeriod=600 Jan 21 18:22:59 crc kubenswrapper[4808]: E0121 18:22:59.657822 4808 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd76d1c92_28d2_4476_8af9_d95cc779425e.slice/crio-3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1.scope\": RecentStats: unable to find data in memory cache]" Jan 21 18:22:59 crc kubenswrapper[4808]: E0121 18:22:59.720954 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:23:00 crc kubenswrapper[4808]: I0121 18:23:00.537354 4808 generic.go:334] "Generic (PLEG): container finished" podID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" exitCode=0 Jan 21 18:23:00 crc kubenswrapper[4808]: I0121 18:23:00.537401 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerDied","Data":"3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1"} Jan 21 18:23:00 crc kubenswrapper[4808]: I0121 18:23:00.537473 4808 scope.go:117] "RemoveContainer" containerID="fdbd789ae8672bd0f6357c9d20a4c92b6be5b8f6f3df753f537861c9d45d3858" Jan 21 18:23:00 crc kubenswrapper[4808]: I0121 18:23:00.538891 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:23:00 crc kubenswrapper[4808]: E0121 18:23:00.539780 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:23:14 crc kubenswrapper[4808]: I0121 18:23:14.520729 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:23:14 crc kubenswrapper[4808]: E0121 18:23:14.521538 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:23:27 crc kubenswrapper[4808]: I0121 18:23:27.520651 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:23:27 crc kubenswrapper[4808]: E0121 18:23:27.521557 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:23:41 crc kubenswrapper[4808]: I0121 18:23:41.520666 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:23:41 crc kubenswrapper[4808]: E0121 18:23:41.521372 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:23:56 crc kubenswrapper[4808]: I0121 18:23:56.521400 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:23:56 crc kubenswrapper[4808]: E0121 18:23:56.522666 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:24:07 crc kubenswrapper[4808]: I0121 18:24:07.521782 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:24:07 crc kubenswrapper[4808]: E0121 18:24:07.525586 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:24:22 crc kubenswrapper[4808]: I0121 18:24:22.520872 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:24:22 crc kubenswrapper[4808]: E0121 18:24:22.521718 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:24:36 crc kubenswrapper[4808]: I0121 18:24:36.520623 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:24:36 crc kubenswrapper[4808]: E0121 18:24:36.521220 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:24:49 crc kubenswrapper[4808]: I0121 18:24:49.520770 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:24:49 crc kubenswrapper[4808]: E0121 18:24:49.521501 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:25:03 crc kubenswrapper[4808]: I0121 18:25:03.521365 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:25:03 crc kubenswrapper[4808]: E0121 18:25:03.522061 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:25:17 crc kubenswrapper[4808]: I0121 18:25:17.524574 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:25:17 crc kubenswrapper[4808]: E0121 18:25:17.525517 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:25:28 crc kubenswrapper[4808]: I0121 18:25:28.520793 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:25:28 crc kubenswrapper[4808]: E0121 18:25:28.521433 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:25:42 crc kubenswrapper[4808]: I0121 18:25:42.520161 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:25:42 crc kubenswrapper[4808]: E0121 18:25:42.521036 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:25:53 crc kubenswrapper[4808]: I0121 18:25:53.520534 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:25:53 crc kubenswrapper[4808]: E0121 18:25:53.521346 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:26:07 crc kubenswrapper[4808]: I0121 18:26:07.522219 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:26:07 crc kubenswrapper[4808]: E0121 18:26:07.523062 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:26:19 crc kubenswrapper[4808]: I0121 18:26:19.521030 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:26:19 crc kubenswrapper[4808]: E0121 18:26:19.521826 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:26:30 crc kubenswrapper[4808]: I0121 18:26:30.521109 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:26:30 crc kubenswrapper[4808]: E0121 18:26:30.521865 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:26:41 crc kubenswrapper[4808]: I0121 18:26:41.520778 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:26:41 crc kubenswrapper[4808]: E0121 18:26:41.521488 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:26:52 crc kubenswrapper[4808]: I0121 18:26:52.521205 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:26:52 crc kubenswrapper[4808]: E0121 18:26:52.523262 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:27:07 crc kubenswrapper[4808]: I0121 18:27:07.521545 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:27:07 crc kubenswrapper[4808]: E0121 18:27:07.523328 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:27:22 crc kubenswrapper[4808]: I0121 18:27:22.520228 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:27:22 crc kubenswrapper[4808]: E0121 18:27:22.521457 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:27:36 crc kubenswrapper[4808]: I0121 18:27:36.520972 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:27:36 crc kubenswrapper[4808]: E0121 18:27:36.522049 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:27:50 crc kubenswrapper[4808]: I0121 18:27:50.522220 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:27:50 crc kubenswrapper[4808]: E0121 18:27:50.523204 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:28:01 crc kubenswrapper[4808]: I0121 18:28:01.521537 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:28:01 crc kubenswrapper[4808]: I0121 18:28:01.760933 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerStarted","Data":"dd13afd7fd43a6e8f7732443fd57a881906abc4ac14c52babeb4773a1dc24aa8"} Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.151417 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483670-bprkf"] Jan 21 18:30:00 crc kubenswrapper[4808]: E0121 18:30:00.152308 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="928a038e-621a-4f7c-822e-cc3b50799fe5" containerName="extract-utilities" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.152324 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="928a038e-621a-4f7c-822e-cc3b50799fe5" containerName="extract-utilities" Jan 21 18:30:00 crc kubenswrapper[4808]: E0121 18:30:00.152344 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e185876e-8a0f-4583-ab42-6325279d31d2" containerName="extract-content" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.152351 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e185876e-8a0f-4583-ab42-6325279d31d2" containerName="extract-content" Jan 21 18:30:00 crc kubenswrapper[4808]: E0121 18:30:00.152365 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="928a038e-621a-4f7c-822e-cc3b50799fe5" containerName="registry-server" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.152372 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="928a038e-621a-4f7c-822e-cc3b50799fe5" containerName="registry-server" Jan 21 18:30:00 crc kubenswrapper[4808]: E0121 18:30:00.152386 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e185876e-8a0f-4583-ab42-6325279d31d2" containerName="registry-server" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.152393 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e185876e-8a0f-4583-ab42-6325279d31d2" containerName="registry-server" Jan 21 18:30:00 crc kubenswrapper[4808]: E0121 18:30:00.152413 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="928a038e-621a-4f7c-822e-cc3b50799fe5" containerName="extract-content" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.152421 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="928a038e-621a-4f7c-822e-cc3b50799fe5" containerName="extract-content" Jan 21 18:30:00 crc kubenswrapper[4808]: E0121 18:30:00.152432 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e185876e-8a0f-4583-ab42-6325279d31d2" containerName="extract-utilities" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.152440 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e185876e-8a0f-4583-ab42-6325279d31d2" containerName="extract-utilities" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.152602 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="928a038e-621a-4f7c-822e-cc3b50799fe5" containerName="registry-server" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.152623 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e185876e-8a0f-4583-ab42-6325279d31d2" containerName="registry-server" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.153149 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483670-bprkf" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.156539 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.156551 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.165470 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483670-bprkf"] Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.313302 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/18af29a4-0c26-41fa-8468-dda82d27de54-config-volume\") pod \"collect-profiles-29483670-bprkf\" (UID: \"18af29a4-0c26-41fa-8468-dda82d27de54\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483670-bprkf" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.313354 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/18af29a4-0c26-41fa-8468-dda82d27de54-secret-volume\") pod \"collect-profiles-29483670-bprkf\" (UID: \"18af29a4-0c26-41fa-8468-dda82d27de54\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483670-bprkf" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.313408 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6j6mf\" (UniqueName: \"kubernetes.io/projected/18af29a4-0c26-41fa-8468-dda82d27de54-kube-api-access-6j6mf\") pod \"collect-profiles-29483670-bprkf\" (UID: \"18af29a4-0c26-41fa-8468-dda82d27de54\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483670-bprkf" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.414756 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/18af29a4-0c26-41fa-8468-dda82d27de54-config-volume\") pod \"collect-profiles-29483670-bprkf\" (UID: \"18af29a4-0c26-41fa-8468-dda82d27de54\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483670-bprkf" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.415136 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/18af29a4-0c26-41fa-8468-dda82d27de54-secret-volume\") pod \"collect-profiles-29483670-bprkf\" (UID: \"18af29a4-0c26-41fa-8468-dda82d27de54\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483670-bprkf" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.415205 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6j6mf\" (UniqueName: \"kubernetes.io/projected/18af29a4-0c26-41fa-8468-dda82d27de54-kube-api-access-6j6mf\") pod \"collect-profiles-29483670-bprkf\" (UID: \"18af29a4-0c26-41fa-8468-dda82d27de54\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483670-bprkf" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.415720 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/18af29a4-0c26-41fa-8468-dda82d27de54-config-volume\") pod \"collect-profiles-29483670-bprkf\" (UID: \"18af29a4-0c26-41fa-8468-dda82d27de54\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483670-bprkf" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.423079 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/18af29a4-0c26-41fa-8468-dda82d27de54-secret-volume\") pod \"collect-profiles-29483670-bprkf\" (UID: \"18af29a4-0c26-41fa-8468-dda82d27de54\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483670-bprkf" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.445356 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6j6mf\" (UniqueName: \"kubernetes.io/projected/18af29a4-0c26-41fa-8468-dda82d27de54-kube-api-access-6j6mf\") pod \"collect-profiles-29483670-bprkf\" (UID: \"18af29a4-0c26-41fa-8468-dda82d27de54\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483670-bprkf" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.487712 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483670-bprkf" Jan 21 18:30:00 crc kubenswrapper[4808]: I0121 18:30:00.920439 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483670-bprkf"] Jan 21 18:30:01 crc kubenswrapper[4808]: I0121 18:30:01.556690 4808 generic.go:334] "Generic (PLEG): container finished" podID="18af29a4-0c26-41fa-8468-dda82d27de54" containerID="b91331098840b1e631f29efeca0b4eb6bc1457d859e3cc1d64eb657b707aa8e8" exitCode=0 Jan 21 18:30:01 crc kubenswrapper[4808]: I0121 18:30:01.556964 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483670-bprkf" event={"ID":"18af29a4-0c26-41fa-8468-dda82d27de54","Type":"ContainerDied","Data":"b91331098840b1e631f29efeca0b4eb6bc1457d859e3cc1d64eb657b707aa8e8"} Jan 21 18:30:01 crc kubenswrapper[4808]: I0121 18:30:01.556989 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483670-bprkf" event={"ID":"18af29a4-0c26-41fa-8468-dda82d27de54","Type":"ContainerStarted","Data":"e6770d76fd01f7d0e2fd6c620a8fee229e2e59c06bed3a04acd17ffd6b8661fd"} Jan 21 18:30:02 crc kubenswrapper[4808]: I0121 18:30:02.789337 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483670-bprkf" Jan 21 18:30:02 crc kubenswrapper[4808]: I0121 18:30:02.956181 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/18af29a4-0c26-41fa-8468-dda82d27de54-config-volume\") pod \"18af29a4-0c26-41fa-8468-dda82d27de54\" (UID: \"18af29a4-0c26-41fa-8468-dda82d27de54\") " Jan 21 18:30:02 crc kubenswrapper[4808]: I0121 18:30:02.956478 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6j6mf\" (UniqueName: \"kubernetes.io/projected/18af29a4-0c26-41fa-8468-dda82d27de54-kube-api-access-6j6mf\") pod \"18af29a4-0c26-41fa-8468-dda82d27de54\" (UID: \"18af29a4-0c26-41fa-8468-dda82d27de54\") " Jan 21 18:30:02 crc kubenswrapper[4808]: I0121 18:30:02.956550 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/18af29a4-0c26-41fa-8468-dda82d27de54-secret-volume\") pod \"18af29a4-0c26-41fa-8468-dda82d27de54\" (UID: \"18af29a4-0c26-41fa-8468-dda82d27de54\") " Jan 21 18:30:02 crc kubenswrapper[4808]: I0121 18:30:02.957916 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18af29a4-0c26-41fa-8468-dda82d27de54-config-volume" (OuterVolumeSpecName: "config-volume") pod "18af29a4-0c26-41fa-8468-dda82d27de54" (UID: "18af29a4-0c26-41fa-8468-dda82d27de54"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:30:02 crc kubenswrapper[4808]: I0121 18:30:02.962615 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18af29a4-0c26-41fa-8468-dda82d27de54-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "18af29a4-0c26-41fa-8468-dda82d27de54" (UID: "18af29a4-0c26-41fa-8468-dda82d27de54"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:30:02 crc kubenswrapper[4808]: I0121 18:30:02.964381 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18af29a4-0c26-41fa-8468-dda82d27de54-kube-api-access-6j6mf" (OuterVolumeSpecName: "kube-api-access-6j6mf") pod "18af29a4-0c26-41fa-8468-dda82d27de54" (UID: "18af29a4-0c26-41fa-8468-dda82d27de54"). InnerVolumeSpecName "kube-api-access-6j6mf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:30:03 crc kubenswrapper[4808]: I0121 18:30:03.058568 4808 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/18af29a4-0c26-41fa-8468-dda82d27de54-config-volume\") on node \"crc\" DevicePath \"\"" Jan 21 18:30:03 crc kubenswrapper[4808]: I0121 18:30:03.058958 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6j6mf\" (UniqueName: \"kubernetes.io/projected/18af29a4-0c26-41fa-8468-dda82d27de54-kube-api-access-6j6mf\") on node \"crc\" DevicePath \"\"" Jan 21 18:30:03 crc kubenswrapper[4808]: I0121 18:30:03.058979 4808 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/18af29a4-0c26-41fa-8468-dda82d27de54-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 21 18:30:03 crc kubenswrapper[4808]: I0121 18:30:03.572170 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483670-bprkf" event={"ID":"18af29a4-0c26-41fa-8468-dda82d27de54","Type":"ContainerDied","Data":"e6770d76fd01f7d0e2fd6c620a8fee229e2e59c06bed3a04acd17ffd6b8661fd"} Jan 21 18:30:03 crc kubenswrapper[4808]: I0121 18:30:03.572208 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6770d76fd01f7d0e2fd6c620a8fee229e2e59c06bed3a04acd17ffd6b8661fd" Jan 21 18:30:03 crc kubenswrapper[4808]: I0121 18:30:03.572317 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483670-bprkf" Jan 21 18:30:03 crc kubenswrapper[4808]: I0121 18:30:03.847949 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2"] Jan 21 18:30:03 crc kubenswrapper[4808]: I0121 18:30:03.851825 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483625-lsqb2"] Jan 21 18:30:05 crc kubenswrapper[4808]: I0121 18:30:05.534424 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0b6ae2d-bfaf-4535-9906-31b19d4fe489" path="/var/lib/kubelet/pods/c0b6ae2d-bfaf-4535-9906-31b19d4fe489/volumes" Jan 21 18:30:07 crc kubenswrapper[4808]: I0121 18:30:06.999930 4808 scope.go:117] "RemoveContainer" containerID="38c8bccfb1c4e3a455b6493d4db9ce026347b6cf3b09d5447ebf3fdd0d50db88" Jan 21 18:30:29 crc kubenswrapper[4808]: I0121 18:30:29.599469 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:30:29 crc kubenswrapper[4808]: I0121 18:30:29.600103 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:30:49 crc kubenswrapper[4808]: I0121 18:30:49.663029 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2b4w7"] Jan 21 18:30:49 crc kubenswrapper[4808]: E0121 18:30:49.664093 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18af29a4-0c26-41fa-8468-dda82d27de54" containerName="collect-profiles" Jan 21 18:30:49 crc kubenswrapper[4808]: I0121 18:30:49.664116 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="18af29a4-0c26-41fa-8468-dda82d27de54" containerName="collect-profiles" Jan 21 18:30:49 crc kubenswrapper[4808]: I0121 18:30:49.664335 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="18af29a4-0c26-41fa-8468-dda82d27de54" containerName="collect-profiles" Jan 21 18:30:49 crc kubenswrapper[4808]: I0121 18:30:49.665750 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2b4w7" Jan 21 18:30:49 crc kubenswrapper[4808]: I0121 18:30:49.677669 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2b4w7"] Jan 21 18:30:49 crc kubenswrapper[4808]: I0121 18:30:49.825453 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bthc\" (UniqueName: \"kubernetes.io/projected/99324a45-e304-4297-8d96-590d68b5dab2-kube-api-access-7bthc\") pod \"redhat-operators-2b4w7\" (UID: \"99324a45-e304-4297-8d96-590d68b5dab2\") " pod="openshift-marketplace/redhat-operators-2b4w7" Jan 21 18:30:49 crc kubenswrapper[4808]: I0121 18:30:49.825579 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99324a45-e304-4297-8d96-590d68b5dab2-utilities\") pod \"redhat-operators-2b4w7\" (UID: \"99324a45-e304-4297-8d96-590d68b5dab2\") " pod="openshift-marketplace/redhat-operators-2b4w7" Jan 21 18:30:49 crc kubenswrapper[4808]: I0121 18:30:49.825609 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99324a45-e304-4297-8d96-590d68b5dab2-catalog-content\") pod \"redhat-operators-2b4w7\" (UID: \"99324a45-e304-4297-8d96-590d68b5dab2\") " pod="openshift-marketplace/redhat-operators-2b4w7" Jan 21 18:30:49 crc kubenswrapper[4808]: I0121 18:30:49.926870 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99324a45-e304-4297-8d96-590d68b5dab2-utilities\") pod \"redhat-operators-2b4w7\" (UID: \"99324a45-e304-4297-8d96-590d68b5dab2\") " pod="openshift-marketplace/redhat-operators-2b4w7" Jan 21 18:30:49 crc kubenswrapper[4808]: I0121 18:30:49.926915 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99324a45-e304-4297-8d96-590d68b5dab2-catalog-content\") pod \"redhat-operators-2b4w7\" (UID: \"99324a45-e304-4297-8d96-590d68b5dab2\") " pod="openshift-marketplace/redhat-operators-2b4w7" Jan 21 18:30:49 crc kubenswrapper[4808]: I0121 18:30:49.926960 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bthc\" (UniqueName: \"kubernetes.io/projected/99324a45-e304-4297-8d96-590d68b5dab2-kube-api-access-7bthc\") pod \"redhat-operators-2b4w7\" (UID: \"99324a45-e304-4297-8d96-590d68b5dab2\") " pod="openshift-marketplace/redhat-operators-2b4w7" Jan 21 18:30:49 crc kubenswrapper[4808]: I0121 18:30:49.928187 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99324a45-e304-4297-8d96-590d68b5dab2-utilities\") pod \"redhat-operators-2b4w7\" (UID: \"99324a45-e304-4297-8d96-590d68b5dab2\") " pod="openshift-marketplace/redhat-operators-2b4w7" Jan 21 18:30:49 crc kubenswrapper[4808]: I0121 18:30:49.928271 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99324a45-e304-4297-8d96-590d68b5dab2-catalog-content\") pod \"redhat-operators-2b4w7\" (UID: \"99324a45-e304-4297-8d96-590d68b5dab2\") " pod="openshift-marketplace/redhat-operators-2b4w7" Jan 21 18:30:49 crc kubenswrapper[4808]: I0121 18:30:49.961136 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bthc\" (UniqueName: \"kubernetes.io/projected/99324a45-e304-4297-8d96-590d68b5dab2-kube-api-access-7bthc\") pod \"redhat-operators-2b4w7\" (UID: \"99324a45-e304-4297-8d96-590d68b5dab2\") " pod="openshift-marketplace/redhat-operators-2b4w7" Jan 21 18:30:49 crc kubenswrapper[4808]: I0121 18:30:49.987893 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2b4w7" Jan 21 18:30:50 crc kubenswrapper[4808]: I0121 18:30:50.289476 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2b4w7"] Jan 21 18:30:50 crc kubenswrapper[4808]: I0121 18:30:50.883871 4808 generic.go:334] "Generic (PLEG): container finished" podID="99324a45-e304-4297-8d96-590d68b5dab2" containerID="127e886676d4ed2774c5c51d18912bd7436d696780734aa587e413642ffb8a42" exitCode=0 Jan 21 18:30:50 crc kubenswrapper[4808]: I0121 18:30:50.883934 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b4w7" event={"ID":"99324a45-e304-4297-8d96-590d68b5dab2","Type":"ContainerDied","Data":"127e886676d4ed2774c5c51d18912bd7436d696780734aa587e413642ffb8a42"} Jan 21 18:30:50 crc kubenswrapper[4808]: I0121 18:30:50.884267 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b4w7" event={"ID":"99324a45-e304-4297-8d96-590d68b5dab2","Type":"ContainerStarted","Data":"a8ab82a99bd22e9f12b0ba028283ff5af78dffa33bd9eed4a738d7167b0b3900"} Jan 21 18:30:50 crc kubenswrapper[4808]: I0121 18:30:50.885625 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 18:30:52 crc kubenswrapper[4808]: I0121 18:30:52.902798 4808 generic.go:334] "Generic (PLEG): container finished" podID="99324a45-e304-4297-8d96-590d68b5dab2" containerID="6713fddc4ff857aff4c67a41b05d5232682447f8b585a4da5edff37b0056d467" exitCode=0 Jan 21 18:30:52 crc kubenswrapper[4808]: I0121 18:30:52.902849 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b4w7" event={"ID":"99324a45-e304-4297-8d96-590d68b5dab2","Type":"ContainerDied","Data":"6713fddc4ff857aff4c67a41b05d5232682447f8b585a4da5edff37b0056d467"} Jan 21 18:30:53 crc kubenswrapper[4808]: I0121 18:30:53.924648 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b4w7" event={"ID":"99324a45-e304-4297-8d96-590d68b5dab2","Type":"ContainerStarted","Data":"65a986029b9fe2d1dd662929e80d12f479e67cda63b672a8dae6e1b8e4d39e5a"} Jan 21 18:30:53 crc kubenswrapper[4808]: I0121 18:30:53.954666 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2b4w7" podStartSLOduration=2.330543079 podStartE2EDuration="4.954647197s" podCreationTimestamp="2026-01-21 18:30:49 +0000 UTC" firstStartedPulling="2026-01-21 18:30:50.885371544 +0000 UTC m=+2085.615882429" lastFinishedPulling="2026-01-21 18:30:53.509475662 +0000 UTC m=+2088.239986547" observedRunningTime="2026-01-21 18:30:53.949379648 +0000 UTC m=+2088.679890543" watchObservedRunningTime="2026-01-21 18:30:53.954647197 +0000 UTC m=+2088.685158082" Jan 21 18:30:59 crc kubenswrapper[4808]: I0121 18:30:59.600425 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:30:59 crc kubenswrapper[4808]: I0121 18:30:59.601392 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:30:59 crc kubenswrapper[4808]: I0121 18:30:59.990663 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2b4w7" Jan 21 18:30:59 crc kubenswrapper[4808]: I0121 18:30:59.990752 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2b4w7" Jan 21 18:31:00 crc kubenswrapper[4808]: I0121 18:31:00.031651 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2b4w7" Jan 21 18:31:00 crc kubenswrapper[4808]: I0121 18:31:00.511443 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2b4w7" Jan 21 18:31:00 crc kubenswrapper[4808]: I0121 18:31:00.560153 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2b4w7"] Jan 21 18:31:02 crc kubenswrapper[4808]: I0121 18:31:02.485487 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2b4w7" podUID="99324a45-e304-4297-8d96-590d68b5dab2" containerName="registry-server" containerID="cri-o://65a986029b9fe2d1dd662929e80d12f479e67cda63b672a8dae6e1b8e4d39e5a" gracePeriod=2 Jan 21 18:31:02 crc kubenswrapper[4808]: I0121 18:31:02.890646 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2b4w7" Jan 21 18:31:02 crc kubenswrapper[4808]: I0121 18:31:02.959382 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bthc\" (UniqueName: \"kubernetes.io/projected/99324a45-e304-4297-8d96-590d68b5dab2-kube-api-access-7bthc\") pod \"99324a45-e304-4297-8d96-590d68b5dab2\" (UID: \"99324a45-e304-4297-8d96-590d68b5dab2\") " Jan 21 18:31:02 crc kubenswrapper[4808]: I0121 18:31:02.959456 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99324a45-e304-4297-8d96-590d68b5dab2-catalog-content\") pod \"99324a45-e304-4297-8d96-590d68b5dab2\" (UID: \"99324a45-e304-4297-8d96-590d68b5dab2\") " Jan 21 18:31:02 crc kubenswrapper[4808]: I0121 18:31:02.959492 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99324a45-e304-4297-8d96-590d68b5dab2-utilities\") pod \"99324a45-e304-4297-8d96-590d68b5dab2\" (UID: \"99324a45-e304-4297-8d96-590d68b5dab2\") " Jan 21 18:31:02 crc kubenswrapper[4808]: I0121 18:31:02.960693 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99324a45-e304-4297-8d96-590d68b5dab2-utilities" (OuterVolumeSpecName: "utilities") pod "99324a45-e304-4297-8d96-590d68b5dab2" (UID: "99324a45-e304-4297-8d96-590d68b5dab2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:31:02 crc kubenswrapper[4808]: I0121 18:31:02.965331 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99324a45-e304-4297-8d96-590d68b5dab2-kube-api-access-7bthc" (OuterVolumeSpecName: "kube-api-access-7bthc") pod "99324a45-e304-4297-8d96-590d68b5dab2" (UID: "99324a45-e304-4297-8d96-590d68b5dab2"). InnerVolumeSpecName "kube-api-access-7bthc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:31:03 crc kubenswrapper[4808]: I0121 18:31:03.060702 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bthc\" (UniqueName: \"kubernetes.io/projected/99324a45-e304-4297-8d96-590d68b5dab2-kube-api-access-7bthc\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:03 crc kubenswrapper[4808]: I0121 18:31:03.060740 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99324a45-e304-4297-8d96-590d68b5dab2-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:03 crc kubenswrapper[4808]: I0121 18:31:03.077141 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99324a45-e304-4297-8d96-590d68b5dab2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "99324a45-e304-4297-8d96-590d68b5dab2" (UID: "99324a45-e304-4297-8d96-590d68b5dab2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:31:03 crc kubenswrapper[4808]: I0121 18:31:03.162645 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99324a45-e304-4297-8d96-590d68b5dab2-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:03 crc kubenswrapper[4808]: I0121 18:31:03.493541 4808 generic.go:334] "Generic (PLEG): container finished" podID="99324a45-e304-4297-8d96-590d68b5dab2" containerID="65a986029b9fe2d1dd662929e80d12f479e67cda63b672a8dae6e1b8e4d39e5a" exitCode=0 Jan 21 18:31:03 crc kubenswrapper[4808]: I0121 18:31:03.493591 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b4w7" event={"ID":"99324a45-e304-4297-8d96-590d68b5dab2","Type":"ContainerDied","Data":"65a986029b9fe2d1dd662929e80d12f479e67cda63b672a8dae6e1b8e4d39e5a"} Jan 21 18:31:03 crc kubenswrapper[4808]: I0121 18:31:03.493626 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b4w7" event={"ID":"99324a45-e304-4297-8d96-590d68b5dab2","Type":"ContainerDied","Data":"a8ab82a99bd22e9f12b0ba028283ff5af78dffa33bd9eed4a738d7167b0b3900"} Jan 21 18:31:03 crc kubenswrapper[4808]: I0121 18:31:03.493646 4808 scope.go:117] "RemoveContainer" containerID="65a986029b9fe2d1dd662929e80d12f479e67cda63b672a8dae6e1b8e4d39e5a" Jan 21 18:31:03 crc kubenswrapper[4808]: I0121 18:31:03.493645 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2b4w7" Jan 21 18:31:03 crc kubenswrapper[4808]: I0121 18:31:03.512485 4808 scope.go:117] "RemoveContainer" containerID="6713fddc4ff857aff4c67a41b05d5232682447f8b585a4da5edff37b0056d467" Jan 21 18:31:03 crc kubenswrapper[4808]: I0121 18:31:03.543536 4808 scope.go:117] "RemoveContainer" containerID="127e886676d4ed2774c5c51d18912bd7436d696780734aa587e413642ffb8a42" Jan 21 18:31:03 crc kubenswrapper[4808]: I0121 18:31:03.545031 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2b4w7"] Jan 21 18:31:03 crc kubenswrapper[4808]: I0121 18:31:03.545068 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2b4w7"] Jan 21 18:31:03 crc kubenswrapper[4808]: I0121 18:31:03.561447 4808 scope.go:117] "RemoveContainer" containerID="65a986029b9fe2d1dd662929e80d12f479e67cda63b672a8dae6e1b8e4d39e5a" Jan 21 18:31:03 crc kubenswrapper[4808]: E0121 18:31:03.561902 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65a986029b9fe2d1dd662929e80d12f479e67cda63b672a8dae6e1b8e4d39e5a\": container with ID starting with 65a986029b9fe2d1dd662929e80d12f479e67cda63b672a8dae6e1b8e4d39e5a not found: ID does not exist" containerID="65a986029b9fe2d1dd662929e80d12f479e67cda63b672a8dae6e1b8e4d39e5a" Jan 21 18:31:03 crc kubenswrapper[4808]: I0121 18:31:03.561944 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65a986029b9fe2d1dd662929e80d12f479e67cda63b672a8dae6e1b8e4d39e5a"} err="failed to get container status \"65a986029b9fe2d1dd662929e80d12f479e67cda63b672a8dae6e1b8e4d39e5a\": rpc error: code = NotFound desc = could not find container \"65a986029b9fe2d1dd662929e80d12f479e67cda63b672a8dae6e1b8e4d39e5a\": container with ID starting with 65a986029b9fe2d1dd662929e80d12f479e67cda63b672a8dae6e1b8e4d39e5a not found: ID does not exist" Jan 21 18:31:03 crc kubenswrapper[4808]: I0121 18:31:03.561971 4808 scope.go:117] "RemoveContainer" containerID="6713fddc4ff857aff4c67a41b05d5232682447f8b585a4da5edff37b0056d467" Jan 21 18:31:03 crc kubenswrapper[4808]: E0121 18:31:03.562486 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6713fddc4ff857aff4c67a41b05d5232682447f8b585a4da5edff37b0056d467\": container with ID starting with 6713fddc4ff857aff4c67a41b05d5232682447f8b585a4da5edff37b0056d467 not found: ID does not exist" containerID="6713fddc4ff857aff4c67a41b05d5232682447f8b585a4da5edff37b0056d467" Jan 21 18:31:03 crc kubenswrapper[4808]: I0121 18:31:03.562530 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6713fddc4ff857aff4c67a41b05d5232682447f8b585a4da5edff37b0056d467"} err="failed to get container status \"6713fddc4ff857aff4c67a41b05d5232682447f8b585a4da5edff37b0056d467\": rpc error: code = NotFound desc = could not find container \"6713fddc4ff857aff4c67a41b05d5232682447f8b585a4da5edff37b0056d467\": container with ID starting with 6713fddc4ff857aff4c67a41b05d5232682447f8b585a4da5edff37b0056d467 not found: ID does not exist" Jan 21 18:31:03 crc kubenswrapper[4808]: I0121 18:31:03.562558 4808 scope.go:117] "RemoveContainer" containerID="127e886676d4ed2774c5c51d18912bd7436d696780734aa587e413642ffb8a42" Jan 21 18:31:03 crc kubenswrapper[4808]: E0121 18:31:03.562866 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"127e886676d4ed2774c5c51d18912bd7436d696780734aa587e413642ffb8a42\": container with ID starting with 127e886676d4ed2774c5c51d18912bd7436d696780734aa587e413642ffb8a42 not found: ID does not exist" containerID="127e886676d4ed2774c5c51d18912bd7436d696780734aa587e413642ffb8a42" Jan 21 18:31:03 crc kubenswrapper[4808]: I0121 18:31:03.562899 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"127e886676d4ed2774c5c51d18912bd7436d696780734aa587e413642ffb8a42"} err="failed to get container status \"127e886676d4ed2774c5c51d18912bd7436d696780734aa587e413642ffb8a42\": rpc error: code = NotFound desc = could not find container \"127e886676d4ed2774c5c51d18912bd7436d696780734aa587e413642ffb8a42\": container with ID starting with 127e886676d4ed2774c5c51d18912bd7436d696780734aa587e413642ffb8a42 not found: ID does not exist" Jan 21 18:31:05 crc kubenswrapper[4808]: I0121 18:31:05.528348 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99324a45-e304-4297-8d96-590d68b5dab2" path="/var/lib/kubelet/pods/99324a45-e304-4297-8d96-590d68b5dab2/volumes" Jan 21 18:31:15 crc kubenswrapper[4808]: I0121 18:31:15.768599 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gjvkr"] Jan 21 18:31:15 crc kubenswrapper[4808]: E0121 18:31:15.769391 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99324a45-e304-4297-8d96-590d68b5dab2" containerName="registry-server" Jan 21 18:31:15 crc kubenswrapper[4808]: I0121 18:31:15.769405 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="99324a45-e304-4297-8d96-590d68b5dab2" containerName="registry-server" Jan 21 18:31:15 crc kubenswrapper[4808]: E0121 18:31:15.769420 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99324a45-e304-4297-8d96-590d68b5dab2" containerName="extract-utilities" Jan 21 18:31:15 crc kubenswrapper[4808]: I0121 18:31:15.769426 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="99324a45-e304-4297-8d96-590d68b5dab2" containerName="extract-utilities" Jan 21 18:31:15 crc kubenswrapper[4808]: E0121 18:31:15.769439 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99324a45-e304-4297-8d96-590d68b5dab2" containerName="extract-content" Jan 21 18:31:15 crc kubenswrapper[4808]: I0121 18:31:15.769445 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="99324a45-e304-4297-8d96-590d68b5dab2" containerName="extract-content" Jan 21 18:31:15 crc kubenswrapper[4808]: I0121 18:31:15.769566 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="99324a45-e304-4297-8d96-590d68b5dab2" containerName="registry-server" Jan 21 18:31:15 crc kubenswrapper[4808]: I0121 18:31:15.773555 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gjvkr" Jan 21 18:31:15 crc kubenswrapper[4808]: I0121 18:31:15.788011 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gjvkr"] Jan 21 18:31:15 crc kubenswrapper[4808]: I0121 18:31:15.909761 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d24c9398-31a5-4a4a-9b63-3426cd4626e3-utilities\") pod \"certified-operators-gjvkr\" (UID: \"d24c9398-31a5-4a4a-9b63-3426cd4626e3\") " pod="openshift-marketplace/certified-operators-gjvkr" Jan 21 18:31:15 crc kubenswrapper[4808]: I0121 18:31:15.910133 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79w9n\" (UniqueName: \"kubernetes.io/projected/d24c9398-31a5-4a4a-9b63-3426cd4626e3-kube-api-access-79w9n\") pod \"certified-operators-gjvkr\" (UID: \"d24c9398-31a5-4a4a-9b63-3426cd4626e3\") " pod="openshift-marketplace/certified-operators-gjvkr" Jan 21 18:31:15 crc kubenswrapper[4808]: I0121 18:31:15.910324 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d24c9398-31a5-4a4a-9b63-3426cd4626e3-catalog-content\") pod \"certified-operators-gjvkr\" (UID: \"d24c9398-31a5-4a4a-9b63-3426cd4626e3\") " pod="openshift-marketplace/certified-operators-gjvkr" Jan 21 18:31:16 crc kubenswrapper[4808]: I0121 18:31:16.011128 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d24c9398-31a5-4a4a-9b63-3426cd4626e3-utilities\") pod \"certified-operators-gjvkr\" (UID: \"d24c9398-31a5-4a4a-9b63-3426cd4626e3\") " pod="openshift-marketplace/certified-operators-gjvkr" Jan 21 18:31:16 crc kubenswrapper[4808]: I0121 18:31:16.011184 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79w9n\" (UniqueName: \"kubernetes.io/projected/d24c9398-31a5-4a4a-9b63-3426cd4626e3-kube-api-access-79w9n\") pod \"certified-operators-gjvkr\" (UID: \"d24c9398-31a5-4a4a-9b63-3426cd4626e3\") " pod="openshift-marketplace/certified-operators-gjvkr" Jan 21 18:31:16 crc kubenswrapper[4808]: I0121 18:31:16.011226 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d24c9398-31a5-4a4a-9b63-3426cd4626e3-catalog-content\") pod \"certified-operators-gjvkr\" (UID: \"d24c9398-31a5-4a4a-9b63-3426cd4626e3\") " pod="openshift-marketplace/certified-operators-gjvkr" Jan 21 18:31:16 crc kubenswrapper[4808]: I0121 18:31:16.011753 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d24c9398-31a5-4a4a-9b63-3426cd4626e3-catalog-content\") pod \"certified-operators-gjvkr\" (UID: \"d24c9398-31a5-4a4a-9b63-3426cd4626e3\") " pod="openshift-marketplace/certified-operators-gjvkr" Jan 21 18:31:16 crc kubenswrapper[4808]: I0121 18:31:16.011756 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d24c9398-31a5-4a4a-9b63-3426cd4626e3-utilities\") pod \"certified-operators-gjvkr\" (UID: \"d24c9398-31a5-4a4a-9b63-3426cd4626e3\") " pod="openshift-marketplace/certified-operators-gjvkr" Jan 21 18:31:16 crc kubenswrapper[4808]: I0121 18:31:16.036144 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79w9n\" (UniqueName: \"kubernetes.io/projected/d24c9398-31a5-4a4a-9b63-3426cd4626e3-kube-api-access-79w9n\") pod \"certified-operators-gjvkr\" (UID: \"d24c9398-31a5-4a4a-9b63-3426cd4626e3\") " pod="openshift-marketplace/certified-operators-gjvkr" Jan 21 18:31:16 crc kubenswrapper[4808]: I0121 18:31:16.096850 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gjvkr" Jan 21 18:31:16 crc kubenswrapper[4808]: I0121 18:31:16.388140 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gjvkr"] Jan 21 18:31:16 crc kubenswrapper[4808]: I0121 18:31:16.587603 4808 generic.go:334] "Generic (PLEG): container finished" podID="d24c9398-31a5-4a4a-9b63-3426cd4626e3" containerID="f084ebb35c566a25d9412562e437d8e37943f5c661fa1a7a9c7c163f66368cd8" exitCode=0 Jan 21 18:31:16 crc kubenswrapper[4808]: I0121 18:31:16.587642 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjvkr" event={"ID":"d24c9398-31a5-4a4a-9b63-3426cd4626e3","Type":"ContainerDied","Data":"f084ebb35c566a25d9412562e437d8e37943f5c661fa1a7a9c7c163f66368cd8"} Jan 21 18:31:16 crc kubenswrapper[4808]: I0121 18:31:16.587681 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjvkr" event={"ID":"d24c9398-31a5-4a4a-9b63-3426cd4626e3","Type":"ContainerStarted","Data":"5652ffaf0dfc81d59a6ecf7cbaa3a347c89112ed59b45c75470811e8463e32d6"} Jan 21 18:31:20 crc kubenswrapper[4808]: I0121 18:31:20.613498 4808 generic.go:334] "Generic (PLEG): container finished" podID="d24c9398-31a5-4a4a-9b63-3426cd4626e3" containerID="2ecee2cdfdba357d5e95bc41b1e4df7e8fd21196269a7567c865f425d4b6ff59" exitCode=0 Jan 21 18:31:20 crc kubenswrapper[4808]: I0121 18:31:20.613728 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjvkr" event={"ID":"d24c9398-31a5-4a4a-9b63-3426cd4626e3","Type":"ContainerDied","Data":"2ecee2cdfdba357d5e95bc41b1e4df7e8fd21196269a7567c865f425d4b6ff59"} Jan 21 18:31:21 crc kubenswrapper[4808]: I0121 18:31:21.620986 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjvkr" event={"ID":"d24c9398-31a5-4a4a-9b63-3426cd4626e3","Type":"ContainerStarted","Data":"56691a79717c3af540eb52f96f4ec7c43637deccbb7b08e2b0dacc1eda4eb4ae"} Jan 21 18:31:21 crc kubenswrapper[4808]: I0121 18:31:21.648706 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gjvkr" podStartSLOduration=2.192399265 podStartE2EDuration="6.648688858s" podCreationTimestamp="2026-01-21 18:31:15 +0000 UTC" firstStartedPulling="2026-01-21 18:31:16.588820451 +0000 UTC m=+2111.319331336" lastFinishedPulling="2026-01-21 18:31:21.045110024 +0000 UTC m=+2115.775620929" observedRunningTime="2026-01-21 18:31:21.645715635 +0000 UTC m=+2116.376226530" watchObservedRunningTime="2026-01-21 18:31:21.648688858 +0000 UTC m=+2116.379199743" Jan 21 18:31:26 crc kubenswrapper[4808]: I0121 18:31:26.097894 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gjvkr" Jan 21 18:31:26 crc kubenswrapper[4808]: I0121 18:31:26.098400 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gjvkr" Jan 21 18:31:26 crc kubenswrapper[4808]: I0121 18:31:26.144356 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gjvkr" Jan 21 18:31:26 crc kubenswrapper[4808]: I0121 18:31:26.715829 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gjvkr" Jan 21 18:31:26 crc kubenswrapper[4808]: I0121 18:31:26.792624 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gjvkr"] Jan 21 18:31:26 crc kubenswrapper[4808]: I0121 18:31:26.847705 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4pfvs"] Jan 21 18:31:26 crc kubenswrapper[4808]: I0121 18:31:26.848123 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4pfvs" podUID="9d1c03b0-e10f-47ae-841c-153e4ae0f9cd" containerName="registry-server" containerID="cri-o://7d1abbc03aa42016ed07b245e1a1ec9e616b4f93dca34fc1da743e210928b23b" gracePeriod=2 Jan 21 18:31:26 crc kubenswrapper[4808]: I0121 18:31:26.857293 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5c9xs"] Jan 21 18:31:26 crc kubenswrapper[4808]: I0121 18:31:26.857689 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5c9xs" podUID="21ebff69-0bfe-427d-8330-4e454159dd01" containerName="registry-server" containerID="cri-o://1838d953c319f8dde63af6b7d6e8d967718bfa31750f291021a420a8ced264b8" gracePeriod=2 Jan 21 18:31:26 crc kubenswrapper[4808]: I0121 18:31:26.863890 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6m94d"] Jan 21 18:31:26 crc kubenswrapper[4808]: I0121 18:31:26.864462 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6m94d" podUID="62cdb877-c15b-4990-8820-3dcc6ffc0b04" containerName="registry-server" containerID="cri-o://adf91b787ad2ac75272151d18611c4d9cbaf09e66530001a605d177fa709ff7c" gracePeriod=2 Jan 21 18:31:26 crc kubenswrapper[4808]: I0121 18:31:26.985942 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cbwpb"] Jan 21 18:31:26 crc kubenswrapper[4808]: I0121 18:31:26.986365 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cbwpb" podUID="1df03daa-d900-4a6d-9776-26f3b05843ce" containerName="registry-server" containerID="cri-o://fa7c5bf063b9350cd7c662bbd4575389c72037d13e2efddf22bc2e186973b0cb" gracePeriod=2 Jan 21 18:31:27 crc kubenswrapper[4808]: I0121 18:31:27.179080 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hzsq2"] Jan 21 18:31:27 crc kubenswrapper[4808]: I0121 18:31:27.179331 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hzsq2" podUID="7f554b03-f857-4309-ba4c-91e8c607a14d" containerName="registry-server" containerID="cri-o://ba1dbe7c160275afae696b07a62f00b7eb61a955782adf02caefbd95939b1493" gracePeriod=2 Jan 21 18:31:27 crc kubenswrapper[4808]: I0121 18:31:27.380580 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kd6dv"] Jan 21 18:31:27 crc kubenswrapper[4808]: I0121 18:31:27.380989 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kd6dv" podUID="dde2d0c2-70a4-4d0a-bc7a-29fd72c50095" containerName="registry-server" containerID="cri-o://86efaa6b013918beeef0b0844278cc5c3993935af964fe0cf6acf0db50a968f8" gracePeriod=2 Jan 21 18:31:27 crc kubenswrapper[4808]: I0121 18:31:27.579440 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nqzlp"] Jan 21 18:31:27 crc kubenswrapper[4808]: I0121 18:31:27.580073 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nqzlp" podUID="c5322a4c-c457-48c5-9c8d-6ce242b4b87b" containerName="registry-server" containerID="cri-o://48bce33ad57d52f0f446058cfd1ba8b2673b597e34eed6ca33bb88dda0af606c" gracePeriod=2 Jan 21 18:31:27 crc kubenswrapper[4808]: I0121 18:31:27.782145 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xstv7"] Jan 21 18:31:27 crc kubenswrapper[4808]: I0121 18:31:27.783079 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xstv7" podUID="173b6b54-1024-4806-8056-985010126525" containerName="registry-server" containerID="cri-o://af4daed6d8653d6e8524448329a6f2108446a3f501ec5904015487ff2138977f" gracePeriod=2 Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.063935 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nqzlp" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.147954 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hzsq2" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.199855 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5322a4c-c457-48c5-9c8d-6ce242b4b87b-catalog-content\") pod \"c5322a4c-c457-48c5-9c8d-6ce242b4b87b\" (UID: \"c5322a4c-c457-48c5-9c8d-6ce242b4b87b\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.199899 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5322a4c-c457-48c5-9c8d-6ce242b4b87b-utilities\") pod \"c5322a4c-c457-48c5-9c8d-6ce242b4b87b\" (UID: \"c5322a4c-c457-48c5-9c8d-6ce242b4b87b\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.199939 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55m49\" (UniqueName: \"kubernetes.io/projected/c5322a4c-c457-48c5-9c8d-6ce242b4b87b-kube-api-access-55m49\") pod \"c5322a4c-c457-48c5-9c8d-6ce242b4b87b\" (UID: \"c5322a4c-c457-48c5-9c8d-6ce242b4b87b\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.201720 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5322a4c-c457-48c5-9c8d-6ce242b4b87b-utilities" (OuterVolumeSpecName: "utilities") pod "c5322a4c-c457-48c5-9c8d-6ce242b4b87b" (UID: "c5322a4c-c457-48c5-9c8d-6ce242b4b87b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.207558 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5322a4c-c457-48c5-9c8d-6ce242b4b87b-kube-api-access-55m49" (OuterVolumeSpecName: "kube-api-access-55m49") pod "c5322a4c-c457-48c5-9c8d-6ce242b4b87b" (UID: "c5322a4c-c457-48c5-9c8d-6ce242b4b87b"). InnerVolumeSpecName "kube-api-access-55m49". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.255158 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5322a4c-c457-48c5-9c8d-6ce242b4b87b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c5322a4c-c457-48c5-9c8d-6ce242b4b87b" (UID: "c5322a4c-c457-48c5-9c8d-6ce242b4b87b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.301288 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5n7qm\" (UniqueName: \"kubernetes.io/projected/7f554b03-f857-4309-ba4c-91e8c607a14d-kube-api-access-5n7qm\") pod \"7f554b03-f857-4309-ba4c-91e8c607a14d\" (UID: \"7f554b03-f857-4309-ba4c-91e8c607a14d\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.301342 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f554b03-f857-4309-ba4c-91e8c607a14d-utilities\") pod \"7f554b03-f857-4309-ba4c-91e8c607a14d\" (UID: \"7f554b03-f857-4309-ba4c-91e8c607a14d\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.301384 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f554b03-f857-4309-ba4c-91e8c607a14d-catalog-content\") pod \"7f554b03-f857-4309-ba4c-91e8c607a14d\" (UID: \"7f554b03-f857-4309-ba4c-91e8c607a14d\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.301623 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5322a4c-c457-48c5-9c8d-6ce242b4b87b-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.301634 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5322a4c-c457-48c5-9c8d-6ce242b4b87b-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.301643 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55m49\" (UniqueName: \"kubernetes.io/projected/c5322a4c-c457-48c5-9c8d-6ce242b4b87b-kube-api-access-55m49\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.303699 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f554b03-f857-4309-ba4c-91e8c607a14d-utilities" (OuterVolumeSpecName: "utilities") pod "7f554b03-f857-4309-ba4c-91e8c607a14d" (UID: "7f554b03-f857-4309-ba4c-91e8c607a14d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.304335 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f554b03-f857-4309-ba4c-91e8c607a14d-kube-api-access-5n7qm" (OuterVolumeSpecName: "kube-api-access-5n7qm") pod "7f554b03-f857-4309-ba4c-91e8c607a14d" (UID: "7f554b03-f857-4309-ba4c-91e8c607a14d"). InnerVolumeSpecName "kube-api-access-5n7qm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.349903 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kd6dv" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.366109 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f554b03-f857-4309-ba4c-91e8c607a14d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7f554b03-f857-4309-ba4c-91e8c607a14d" (UID: "7f554b03-f857-4309-ba4c-91e8c607a14d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.408908 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mbgq\" (UniqueName: \"kubernetes.io/projected/dde2d0c2-70a4-4d0a-bc7a-29fd72c50095-kube-api-access-7mbgq\") pod \"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095\" (UID: \"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.408984 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dde2d0c2-70a4-4d0a-bc7a-29fd72c50095-catalog-content\") pod \"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095\" (UID: \"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.409076 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dde2d0c2-70a4-4d0a-bc7a-29fd72c50095-utilities\") pod \"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095\" (UID: \"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.418699 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dde2d0c2-70a4-4d0a-bc7a-29fd72c50095-utilities" (OuterVolumeSpecName: "utilities") pod "dde2d0c2-70a4-4d0a-bc7a-29fd72c50095" (UID: "dde2d0c2-70a4-4d0a-bc7a-29fd72c50095"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.418932 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5n7qm\" (UniqueName: \"kubernetes.io/projected/7f554b03-f857-4309-ba4c-91e8c607a14d-kube-api-access-5n7qm\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.418950 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f554b03-f857-4309-ba4c-91e8c607a14d-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.418962 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f554b03-f857-4309-ba4c-91e8c607a14d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.418972 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dde2d0c2-70a4-4d0a-bc7a-29fd72c50095-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.419910 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dde2d0c2-70a4-4d0a-bc7a-29fd72c50095-kube-api-access-7mbgq" (OuterVolumeSpecName: "kube-api-access-7mbgq") pod "dde2d0c2-70a4-4d0a-bc7a-29fd72c50095" (UID: "dde2d0c2-70a4-4d0a-bc7a-29fd72c50095"). InnerVolumeSpecName "kube-api-access-7mbgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.456495 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dde2d0c2-70a4-4d0a-bc7a-29fd72c50095-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dde2d0c2-70a4-4d0a-bc7a-29fd72c50095" (UID: "dde2d0c2-70a4-4d0a-bc7a-29fd72c50095"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.501033 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4pfvs" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.510443 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5c9xs" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.519382 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsx8z\" (UniqueName: \"kubernetes.io/projected/9d1c03b0-e10f-47ae-841c-153e4ae0f9cd-kube-api-access-gsx8z\") pod \"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd\" (UID: \"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.519469 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d1c03b0-e10f-47ae-841c-153e4ae0f9cd-utilities\") pod \"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd\" (UID: \"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.519596 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d1c03b0-e10f-47ae-841c-153e4ae0f9cd-catalog-content\") pod \"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd\" (UID: \"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.519858 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mbgq\" (UniqueName: \"kubernetes.io/projected/dde2d0c2-70a4-4d0a-bc7a-29fd72c50095-kube-api-access-7mbgq\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.519872 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dde2d0c2-70a4-4d0a-bc7a-29fd72c50095-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.521587 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d1c03b0-e10f-47ae-841c-153e4ae0f9cd-utilities" (OuterVolumeSpecName: "utilities") pod "9d1c03b0-e10f-47ae-841c-153e4ae0f9cd" (UID: "9d1c03b0-e10f-47ae-841c-153e4ae0f9cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.535087 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d1c03b0-e10f-47ae-841c-153e4ae0f9cd-kube-api-access-gsx8z" (OuterVolumeSpecName: "kube-api-access-gsx8z") pod "9d1c03b0-e10f-47ae-841c-153e4ae0f9cd" (UID: "9d1c03b0-e10f-47ae-841c-153e4ae0f9cd"). InnerVolumeSpecName "kube-api-access-gsx8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.556580 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6m94d" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.619329 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d1c03b0-e10f-47ae-841c-153e4ae0f9cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9d1c03b0-e10f-47ae-841c-153e4ae0f9cd" (UID: "9d1c03b0-e10f-47ae-841c-153e4ae0f9cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.622765 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62cdb877-c15b-4990-8820-3dcc6ffc0b04-utilities\") pod \"62cdb877-c15b-4990-8820-3dcc6ffc0b04\" (UID: \"62cdb877-c15b-4990-8820-3dcc6ffc0b04\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.622882 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p69s2\" (UniqueName: \"kubernetes.io/projected/21ebff69-0bfe-427d-8330-4e454159dd01-kube-api-access-p69s2\") pod \"21ebff69-0bfe-427d-8330-4e454159dd01\" (UID: \"21ebff69-0bfe-427d-8330-4e454159dd01\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.622949 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62cdb877-c15b-4990-8820-3dcc6ffc0b04-catalog-content\") pod \"62cdb877-c15b-4990-8820-3dcc6ffc0b04\" (UID: \"62cdb877-c15b-4990-8820-3dcc6ffc0b04\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.622965 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8snm\" (UniqueName: \"kubernetes.io/projected/62cdb877-c15b-4990-8820-3dcc6ffc0b04-kube-api-access-v8snm\") pod \"62cdb877-c15b-4990-8820-3dcc6ffc0b04\" (UID: \"62cdb877-c15b-4990-8820-3dcc6ffc0b04\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.622993 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21ebff69-0bfe-427d-8330-4e454159dd01-catalog-content\") pod \"21ebff69-0bfe-427d-8330-4e454159dd01\" (UID: \"21ebff69-0bfe-427d-8330-4e454159dd01\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.623018 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21ebff69-0bfe-427d-8330-4e454159dd01-utilities\") pod \"21ebff69-0bfe-427d-8330-4e454159dd01\" (UID: \"21ebff69-0bfe-427d-8330-4e454159dd01\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.623284 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d1c03b0-e10f-47ae-841c-153e4ae0f9cd-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.623295 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsx8z\" (UniqueName: \"kubernetes.io/projected/9d1c03b0-e10f-47ae-841c-153e4ae0f9cd-kube-api-access-gsx8z\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.623305 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d1c03b0-e10f-47ae-841c-153e4ae0f9cd-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.625107 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21ebff69-0bfe-427d-8330-4e454159dd01-utilities" (OuterVolumeSpecName: "utilities") pod "21ebff69-0bfe-427d-8330-4e454159dd01" (UID: "21ebff69-0bfe-427d-8330-4e454159dd01"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.626065 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62cdb877-c15b-4990-8820-3dcc6ffc0b04-utilities" (OuterVolumeSpecName: "utilities") pod "62cdb877-c15b-4990-8820-3dcc6ffc0b04" (UID: "62cdb877-c15b-4990-8820-3dcc6ffc0b04"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.646539 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21ebff69-0bfe-427d-8330-4e454159dd01-kube-api-access-p69s2" (OuterVolumeSpecName: "kube-api-access-p69s2") pod "21ebff69-0bfe-427d-8330-4e454159dd01" (UID: "21ebff69-0bfe-427d-8330-4e454159dd01"). InnerVolumeSpecName "kube-api-access-p69s2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.646652 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62cdb877-c15b-4990-8820-3dcc6ffc0b04-kube-api-access-v8snm" (OuterVolumeSpecName: "kube-api-access-v8snm") pod "62cdb877-c15b-4990-8820-3dcc6ffc0b04" (UID: "62cdb877-c15b-4990-8820-3dcc6ffc0b04"). InnerVolumeSpecName "kube-api-access-v8snm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.672692 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xstv7" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.714062 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62cdb877-c15b-4990-8820-3dcc6ffc0b04-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "62cdb877-c15b-4990-8820-3dcc6ffc0b04" (UID: "62cdb877-c15b-4990-8820-3dcc6ffc0b04"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.714382 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21ebff69-0bfe-427d-8330-4e454159dd01-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "21ebff69-0bfe-427d-8330-4e454159dd01" (UID: "21ebff69-0bfe-427d-8330-4e454159dd01"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.718512 4808 generic.go:334] "Generic (PLEG): container finished" podID="dde2d0c2-70a4-4d0a-bc7a-29fd72c50095" containerID="86efaa6b013918beeef0b0844278cc5c3993935af964fe0cf6acf0db50a968f8" exitCode=0 Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.718576 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kd6dv" event={"ID":"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095","Type":"ContainerDied","Data":"86efaa6b013918beeef0b0844278cc5c3993935af964fe0cf6acf0db50a968f8"} Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.718606 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kd6dv" event={"ID":"dde2d0c2-70a4-4d0a-bc7a-29fd72c50095","Type":"ContainerDied","Data":"be07c45071d3263924afee213f2475c067cbb0b5ca5af893e19b86b606764c78"} Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.718623 4808 scope.go:117] "RemoveContainer" containerID="86efaa6b013918beeef0b0844278cc5c3993935af964fe0cf6acf0db50a968f8" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.718764 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kd6dv" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.723176 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cbwpb" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.724681 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ds56w\" (UniqueName: \"kubernetes.io/projected/173b6b54-1024-4806-8056-985010126525-kube-api-access-ds56w\") pod \"173b6b54-1024-4806-8056-985010126525\" (UID: \"173b6b54-1024-4806-8056-985010126525\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.724836 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/173b6b54-1024-4806-8056-985010126525-utilities\") pod \"173b6b54-1024-4806-8056-985010126525\" (UID: \"173b6b54-1024-4806-8056-985010126525\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.724883 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/173b6b54-1024-4806-8056-985010126525-catalog-content\") pod \"173b6b54-1024-4806-8056-985010126525\" (UID: \"173b6b54-1024-4806-8056-985010126525\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.725958 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/173b6b54-1024-4806-8056-985010126525-utilities" (OuterVolumeSpecName: "utilities") pod "173b6b54-1024-4806-8056-985010126525" (UID: "173b6b54-1024-4806-8056-985010126525"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.726732 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62cdb877-c15b-4990-8820-3dcc6ffc0b04-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.726758 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8snm\" (UniqueName: \"kubernetes.io/projected/62cdb877-c15b-4990-8820-3dcc6ffc0b04-kube-api-access-v8snm\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.729326 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21ebff69-0bfe-427d-8330-4e454159dd01-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.729340 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21ebff69-0bfe-427d-8330-4e454159dd01-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.729349 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/173b6b54-1024-4806-8056-985010126525-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.729359 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62cdb877-c15b-4990-8820-3dcc6ffc0b04-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.729368 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p69s2\" (UniqueName: \"kubernetes.io/projected/21ebff69-0bfe-427d-8330-4e454159dd01-kube-api-access-p69s2\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.727234 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5c9xs" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.729423 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/173b6b54-1024-4806-8056-985010126525-kube-api-access-ds56w" (OuterVolumeSpecName: "kube-api-access-ds56w") pod "173b6b54-1024-4806-8056-985010126525" (UID: "173b6b54-1024-4806-8056-985010126525"). InnerVolumeSpecName "kube-api-access-ds56w". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.726972 4808 generic.go:334] "Generic (PLEG): container finished" podID="21ebff69-0bfe-427d-8330-4e454159dd01" containerID="1838d953c319f8dde63af6b7d6e8d967718bfa31750f291021a420a8ced264b8" exitCode=0 Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.727150 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5c9xs" event={"ID":"21ebff69-0bfe-427d-8330-4e454159dd01","Type":"ContainerDied","Data":"1838d953c319f8dde63af6b7d6e8d967718bfa31750f291021a420a8ced264b8"} Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.729544 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5c9xs" event={"ID":"21ebff69-0bfe-427d-8330-4e454159dd01","Type":"ContainerDied","Data":"9040772e5d46bfead7cad29d4879c89e8266e7687fc02b14c492e602792ebf88"} Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.736003 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xstv7" event={"ID":"173b6b54-1024-4806-8056-985010126525","Type":"ContainerDied","Data":"af4daed6d8653d6e8524448329a6f2108446a3f501ec5904015487ff2138977f"} Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.735954 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xstv7" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.735775 4808 generic.go:334] "Generic (PLEG): container finished" podID="173b6b54-1024-4806-8056-985010126525" containerID="af4daed6d8653d6e8524448329a6f2108446a3f501ec5904015487ff2138977f" exitCode=0 Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.736821 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xstv7" event={"ID":"173b6b54-1024-4806-8056-985010126525","Type":"ContainerDied","Data":"2b789d2ed58e2ed957ab68a4d09ed6f179d74417c4b7179e1a9c082eb2e0e08c"} Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.750434 4808 generic.go:334] "Generic (PLEG): container finished" podID="7f554b03-f857-4309-ba4c-91e8c607a14d" containerID="ba1dbe7c160275afae696b07a62f00b7eb61a955782adf02caefbd95939b1493" exitCode=0 Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.750510 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hzsq2" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.750515 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hzsq2" event={"ID":"7f554b03-f857-4309-ba4c-91e8c607a14d","Type":"ContainerDied","Data":"ba1dbe7c160275afae696b07a62f00b7eb61a955782adf02caefbd95939b1493"} Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.750544 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hzsq2" event={"ID":"7f554b03-f857-4309-ba4c-91e8c607a14d","Type":"ContainerDied","Data":"31a325e79109c191d7fd701061369dcb0bf66e6bc033bf9b6b7225d61cdb4bd6"} Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.761033 4808 generic.go:334] "Generic (PLEG): container finished" podID="1df03daa-d900-4a6d-9776-26f3b05843ce" containerID="fa7c5bf063b9350cd7c662bbd4575389c72037d13e2efddf22bc2e186973b0cb" exitCode=0 Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.761150 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cbwpb" event={"ID":"1df03daa-d900-4a6d-9776-26f3b05843ce","Type":"ContainerDied","Data":"fa7c5bf063b9350cd7c662bbd4575389c72037d13e2efddf22bc2e186973b0cb"} Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.761283 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cbwpb" event={"ID":"1df03daa-d900-4a6d-9776-26f3b05843ce","Type":"ContainerDied","Data":"341be48dd7e1672a511bd2cbc7dc1fc96983b7b7430fd43553f983fb8b68d9fd"} Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.761850 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cbwpb" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.770955 4808 scope.go:117] "RemoveContainer" containerID="bfdb88542a84b2bb64980b75574f46a75b9be79cd15bf33511410187c75f8198" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.773332 4808 generic.go:334] "Generic (PLEG): container finished" podID="9d1c03b0-e10f-47ae-841c-153e4ae0f9cd" containerID="7d1abbc03aa42016ed07b245e1a1ec9e616b4f93dca34fc1da743e210928b23b" exitCode=0 Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.774216 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4pfvs" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.774361 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pfvs" event={"ID":"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd","Type":"ContainerDied","Data":"7d1abbc03aa42016ed07b245e1a1ec9e616b4f93dca34fc1da743e210928b23b"} Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.776820 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pfvs" event={"ID":"9d1c03b0-e10f-47ae-841c-153e4ae0f9cd","Type":"ContainerDied","Data":"a1bc51b30d252ba93ed5610738e6486733cf5b2409100eed98a5ad1a067f6f29"} Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.794082 4808 generic.go:334] "Generic (PLEG): container finished" podID="c5322a4c-c457-48c5-9c8d-6ce242b4b87b" containerID="48bce33ad57d52f0f446058cfd1ba8b2673b597e34eed6ca33bb88dda0af606c" exitCode=0 Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.794156 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqzlp" event={"ID":"c5322a4c-c457-48c5-9c8d-6ce242b4b87b","Type":"ContainerDied","Data":"48bce33ad57d52f0f446058cfd1ba8b2673b597e34eed6ca33bb88dda0af606c"} Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.794187 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqzlp" event={"ID":"c5322a4c-c457-48c5-9c8d-6ce242b4b87b","Type":"ContainerDied","Data":"b514c2248d867d2c92ce129f2344ff5d506b91de25738057f1f196306a558a92"} Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.794552 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nqzlp" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.801735 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kd6dv"] Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.807863 4808 generic.go:334] "Generic (PLEG): container finished" podID="62cdb877-c15b-4990-8820-3dcc6ffc0b04" containerID="adf91b787ad2ac75272151d18611c4d9cbaf09e66530001a605d177fa709ff7c" exitCode=0 Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.808446 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6m94d" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.808615 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6m94d" event={"ID":"62cdb877-c15b-4990-8820-3dcc6ffc0b04","Type":"ContainerDied","Data":"adf91b787ad2ac75272151d18611c4d9cbaf09e66530001a605d177fa709ff7c"} Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.813797 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kd6dv"] Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.813940 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6m94d" event={"ID":"62cdb877-c15b-4990-8820-3dcc6ffc0b04","Type":"ContainerDied","Data":"b5c91b870a29cb0d04d1dc9edd6b03fe5ebc7f5f85546d433b745bc38c488d4a"} Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.816685 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5c9xs"] Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.821486 4808 scope.go:117] "RemoveContainer" containerID="3d6c6e1a6563688f3d99ece628abdd6b6a6294db7fd5393c3979c7540d78fcc6" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.821683 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5c9xs"] Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.828255 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hzsq2"] Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.830479 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mv79m\" (UniqueName: \"kubernetes.io/projected/1df03daa-d900-4a6d-9776-26f3b05843ce-kube-api-access-mv79m\") pod \"1df03daa-d900-4a6d-9776-26f3b05843ce\" (UID: \"1df03daa-d900-4a6d-9776-26f3b05843ce\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.830534 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df03daa-d900-4a6d-9776-26f3b05843ce-utilities\") pod \"1df03daa-d900-4a6d-9776-26f3b05843ce\" (UID: \"1df03daa-d900-4a6d-9776-26f3b05843ce\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.830581 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df03daa-d900-4a6d-9776-26f3b05843ce-catalog-content\") pod \"1df03daa-d900-4a6d-9776-26f3b05843ce\" (UID: \"1df03daa-d900-4a6d-9776-26f3b05843ce\") " Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.831031 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ds56w\" (UniqueName: \"kubernetes.io/projected/173b6b54-1024-4806-8056-985010126525-kube-api-access-ds56w\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.832449 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hzsq2"] Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.832622 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1df03daa-d900-4a6d-9776-26f3b05843ce-utilities" (OuterVolumeSpecName: "utilities") pod "1df03daa-d900-4a6d-9776-26f3b05843ce" (UID: "1df03daa-d900-4a6d-9776-26f3b05843ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.844317 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4pfvs"] Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.853575 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/173b6b54-1024-4806-8056-985010126525-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "173b6b54-1024-4806-8056-985010126525" (UID: "173b6b54-1024-4806-8056-985010126525"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.853658 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4pfvs"] Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.856030 4808 scope.go:117] "RemoveContainer" containerID="86efaa6b013918beeef0b0844278cc5c3993935af964fe0cf6acf0db50a968f8" Jan 21 18:31:28 crc kubenswrapper[4808]: E0121 18:31:28.856557 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86efaa6b013918beeef0b0844278cc5c3993935af964fe0cf6acf0db50a968f8\": container with ID starting with 86efaa6b013918beeef0b0844278cc5c3993935af964fe0cf6acf0db50a968f8 not found: ID does not exist" containerID="86efaa6b013918beeef0b0844278cc5c3993935af964fe0cf6acf0db50a968f8" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.856658 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86efaa6b013918beeef0b0844278cc5c3993935af964fe0cf6acf0db50a968f8"} err="failed to get container status \"86efaa6b013918beeef0b0844278cc5c3993935af964fe0cf6acf0db50a968f8\": rpc error: code = NotFound desc = could not find container \"86efaa6b013918beeef0b0844278cc5c3993935af964fe0cf6acf0db50a968f8\": container with ID starting with 86efaa6b013918beeef0b0844278cc5c3993935af964fe0cf6acf0db50a968f8 not found: ID does not exist" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.856739 4808 scope.go:117] "RemoveContainer" containerID="bfdb88542a84b2bb64980b75574f46a75b9be79cd15bf33511410187c75f8198" Jan 21 18:31:28 crc kubenswrapper[4808]: E0121 18:31:28.857197 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfdb88542a84b2bb64980b75574f46a75b9be79cd15bf33511410187c75f8198\": container with ID starting with bfdb88542a84b2bb64980b75574f46a75b9be79cd15bf33511410187c75f8198 not found: ID does not exist" containerID="bfdb88542a84b2bb64980b75574f46a75b9be79cd15bf33511410187c75f8198" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.857273 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfdb88542a84b2bb64980b75574f46a75b9be79cd15bf33511410187c75f8198"} err="failed to get container status \"bfdb88542a84b2bb64980b75574f46a75b9be79cd15bf33511410187c75f8198\": rpc error: code = NotFound desc = could not find container \"bfdb88542a84b2bb64980b75574f46a75b9be79cd15bf33511410187c75f8198\": container with ID starting with bfdb88542a84b2bb64980b75574f46a75b9be79cd15bf33511410187c75f8198 not found: ID does not exist" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.857302 4808 scope.go:117] "RemoveContainer" containerID="3d6c6e1a6563688f3d99ece628abdd6b6a6294db7fd5393c3979c7540d78fcc6" Jan 21 18:31:28 crc kubenswrapper[4808]: E0121 18:31:28.857693 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d6c6e1a6563688f3d99ece628abdd6b6a6294db7fd5393c3979c7540d78fcc6\": container with ID starting with 3d6c6e1a6563688f3d99ece628abdd6b6a6294db7fd5393c3979c7540d78fcc6 not found: ID does not exist" containerID="3d6c6e1a6563688f3d99ece628abdd6b6a6294db7fd5393c3979c7540d78fcc6" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.857769 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d6c6e1a6563688f3d99ece628abdd6b6a6294db7fd5393c3979c7540d78fcc6"} err="failed to get container status \"3d6c6e1a6563688f3d99ece628abdd6b6a6294db7fd5393c3979c7540d78fcc6\": rpc error: code = NotFound desc = could not find container \"3d6c6e1a6563688f3d99ece628abdd6b6a6294db7fd5393c3979c7540d78fcc6\": container with ID starting with 3d6c6e1a6563688f3d99ece628abdd6b6a6294db7fd5393c3979c7540d78fcc6 not found: ID does not exist" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.857843 4808 scope.go:117] "RemoveContainer" containerID="1838d953c319f8dde63af6b7d6e8d967718bfa31750f291021a420a8ced264b8" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.864004 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nqzlp"] Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.872536 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1df03daa-d900-4a6d-9776-26f3b05843ce-kube-api-access-mv79m" (OuterVolumeSpecName: "kube-api-access-mv79m") pod "1df03daa-d900-4a6d-9776-26f3b05843ce" (UID: "1df03daa-d900-4a6d-9776-26f3b05843ce"). InnerVolumeSpecName "kube-api-access-mv79m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.884693 4808 scope.go:117] "RemoveContainer" containerID="a6605e09313eafff3dbcd22afed6ee2336c16ae8ec7853efc4be5ee96817229c" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.885905 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nqzlp"] Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.895408 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6m94d"] Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.899187 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6m94d"] Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.902368 4808 scope.go:117] "RemoveContainer" containerID="165fbd2262e741d57cd375773c251eb17b3fa4b02ee5a6658ab73a6334f6108e" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.921077 4808 scope.go:117] "RemoveContainer" containerID="1838d953c319f8dde63af6b7d6e8d967718bfa31750f291021a420a8ced264b8" Jan 21 18:31:28 crc kubenswrapper[4808]: E0121 18:31:28.923365 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1838d953c319f8dde63af6b7d6e8d967718bfa31750f291021a420a8ced264b8\": container with ID starting with 1838d953c319f8dde63af6b7d6e8d967718bfa31750f291021a420a8ced264b8 not found: ID does not exist" containerID="1838d953c319f8dde63af6b7d6e8d967718bfa31750f291021a420a8ced264b8" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.923397 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1838d953c319f8dde63af6b7d6e8d967718bfa31750f291021a420a8ced264b8"} err="failed to get container status \"1838d953c319f8dde63af6b7d6e8d967718bfa31750f291021a420a8ced264b8\": rpc error: code = NotFound desc = could not find container \"1838d953c319f8dde63af6b7d6e8d967718bfa31750f291021a420a8ced264b8\": container with ID starting with 1838d953c319f8dde63af6b7d6e8d967718bfa31750f291021a420a8ced264b8 not found: ID does not exist" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.923435 4808 scope.go:117] "RemoveContainer" containerID="a6605e09313eafff3dbcd22afed6ee2336c16ae8ec7853efc4be5ee96817229c" Jan 21 18:31:28 crc kubenswrapper[4808]: E0121 18:31:28.924207 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6605e09313eafff3dbcd22afed6ee2336c16ae8ec7853efc4be5ee96817229c\": container with ID starting with a6605e09313eafff3dbcd22afed6ee2336c16ae8ec7853efc4be5ee96817229c not found: ID does not exist" containerID="a6605e09313eafff3dbcd22afed6ee2336c16ae8ec7853efc4be5ee96817229c" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.924300 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6605e09313eafff3dbcd22afed6ee2336c16ae8ec7853efc4be5ee96817229c"} err="failed to get container status \"a6605e09313eafff3dbcd22afed6ee2336c16ae8ec7853efc4be5ee96817229c\": rpc error: code = NotFound desc = could not find container \"a6605e09313eafff3dbcd22afed6ee2336c16ae8ec7853efc4be5ee96817229c\": container with ID starting with a6605e09313eafff3dbcd22afed6ee2336c16ae8ec7853efc4be5ee96817229c not found: ID does not exist" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.924333 4808 scope.go:117] "RemoveContainer" containerID="165fbd2262e741d57cd375773c251eb17b3fa4b02ee5a6658ab73a6334f6108e" Jan 21 18:31:28 crc kubenswrapper[4808]: E0121 18:31:28.927538 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"165fbd2262e741d57cd375773c251eb17b3fa4b02ee5a6658ab73a6334f6108e\": container with ID starting with 165fbd2262e741d57cd375773c251eb17b3fa4b02ee5a6658ab73a6334f6108e not found: ID does not exist" containerID="165fbd2262e741d57cd375773c251eb17b3fa4b02ee5a6658ab73a6334f6108e" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.927567 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"165fbd2262e741d57cd375773c251eb17b3fa4b02ee5a6658ab73a6334f6108e"} err="failed to get container status \"165fbd2262e741d57cd375773c251eb17b3fa4b02ee5a6658ab73a6334f6108e\": rpc error: code = NotFound desc = could not find container \"165fbd2262e741d57cd375773c251eb17b3fa4b02ee5a6658ab73a6334f6108e\": container with ID starting with 165fbd2262e741d57cd375773c251eb17b3fa4b02ee5a6658ab73a6334f6108e not found: ID does not exist" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.927588 4808 scope.go:117] "RemoveContainer" containerID="af4daed6d8653d6e8524448329a6f2108446a3f501ec5904015487ff2138977f" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.928038 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1df03daa-d900-4a6d-9776-26f3b05843ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1df03daa-d900-4a6d-9776-26f3b05843ce" (UID: "1df03daa-d900-4a6d-9776-26f3b05843ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.933041 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/173b6b54-1024-4806-8056-985010126525-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.933078 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mv79m\" (UniqueName: \"kubernetes.io/projected/1df03daa-d900-4a6d-9776-26f3b05843ce-kube-api-access-mv79m\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.933094 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df03daa-d900-4a6d-9776-26f3b05843ce-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.933106 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df03daa-d900-4a6d-9776-26f3b05843ce-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.950606 4808 scope.go:117] "RemoveContainer" containerID="5202c43b50f1265bbd8a3ed0aa836614a8f0ddc4a82081f921e026f26dd76b02" Jan 21 18:31:28 crc kubenswrapper[4808]: I0121 18:31:28.976684 4808 scope.go:117] "RemoveContainer" containerID="1718c9d40462fd7d04dc435d2d6e3b79508ec08e1baad202f6eae85e056b2708" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.021996 4808 scope.go:117] "RemoveContainer" containerID="af4daed6d8653d6e8524448329a6f2108446a3f501ec5904015487ff2138977f" Jan 21 18:31:29 crc kubenswrapper[4808]: E0121 18:31:29.022531 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af4daed6d8653d6e8524448329a6f2108446a3f501ec5904015487ff2138977f\": container with ID starting with af4daed6d8653d6e8524448329a6f2108446a3f501ec5904015487ff2138977f not found: ID does not exist" containerID="af4daed6d8653d6e8524448329a6f2108446a3f501ec5904015487ff2138977f" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.022662 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af4daed6d8653d6e8524448329a6f2108446a3f501ec5904015487ff2138977f"} err="failed to get container status \"af4daed6d8653d6e8524448329a6f2108446a3f501ec5904015487ff2138977f\": rpc error: code = NotFound desc = could not find container \"af4daed6d8653d6e8524448329a6f2108446a3f501ec5904015487ff2138977f\": container with ID starting with af4daed6d8653d6e8524448329a6f2108446a3f501ec5904015487ff2138977f not found: ID does not exist" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.022752 4808 scope.go:117] "RemoveContainer" containerID="5202c43b50f1265bbd8a3ed0aa836614a8f0ddc4a82081f921e026f26dd76b02" Jan 21 18:31:29 crc kubenswrapper[4808]: E0121 18:31:29.023075 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5202c43b50f1265bbd8a3ed0aa836614a8f0ddc4a82081f921e026f26dd76b02\": container with ID starting with 5202c43b50f1265bbd8a3ed0aa836614a8f0ddc4a82081f921e026f26dd76b02 not found: ID does not exist" containerID="5202c43b50f1265bbd8a3ed0aa836614a8f0ddc4a82081f921e026f26dd76b02" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.023111 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5202c43b50f1265bbd8a3ed0aa836614a8f0ddc4a82081f921e026f26dd76b02"} err="failed to get container status \"5202c43b50f1265bbd8a3ed0aa836614a8f0ddc4a82081f921e026f26dd76b02\": rpc error: code = NotFound desc = could not find container \"5202c43b50f1265bbd8a3ed0aa836614a8f0ddc4a82081f921e026f26dd76b02\": container with ID starting with 5202c43b50f1265bbd8a3ed0aa836614a8f0ddc4a82081f921e026f26dd76b02 not found: ID does not exist" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.023137 4808 scope.go:117] "RemoveContainer" containerID="1718c9d40462fd7d04dc435d2d6e3b79508ec08e1baad202f6eae85e056b2708" Jan 21 18:31:29 crc kubenswrapper[4808]: E0121 18:31:29.023350 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1718c9d40462fd7d04dc435d2d6e3b79508ec08e1baad202f6eae85e056b2708\": container with ID starting with 1718c9d40462fd7d04dc435d2d6e3b79508ec08e1baad202f6eae85e056b2708 not found: ID does not exist" containerID="1718c9d40462fd7d04dc435d2d6e3b79508ec08e1baad202f6eae85e056b2708" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.023379 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1718c9d40462fd7d04dc435d2d6e3b79508ec08e1baad202f6eae85e056b2708"} err="failed to get container status \"1718c9d40462fd7d04dc435d2d6e3b79508ec08e1baad202f6eae85e056b2708\": rpc error: code = NotFound desc = could not find container \"1718c9d40462fd7d04dc435d2d6e3b79508ec08e1baad202f6eae85e056b2708\": container with ID starting with 1718c9d40462fd7d04dc435d2d6e3b79508ec08e1baad202f6eae85e056b2708 not found: ID does not exist" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.023395 4808 scope.go:117] "RemoveContainer" containerID="ba1dbe7c160275afae696b07a62f00b7eb61a955782adf02caefbd95939b1493" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.055584 4808 scope.go:117] "RemoveContainer" containerID="6624bc1e34644ec644da76e9ba8e3fedf50bd4481db75a8fc10f7b987cf03d4a" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.085871 4808 scope.go:117] "RemoveContainer" containerID="1c2540cfd9f6f2aa5e9c8226454898bf6f7394cd1609b32776dbc9ddc106fc74" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.096970 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xstv7"] Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.118881 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xstv7"] Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.123560 4808 scope.go:117] "RemoveContainer" containerID="ba1dbe7c160275afae696b07a62f00b7eb61a955782adf02caefbd95939b1493" Jan 21 18:31:29 crc kubenswrapper[4808]: E0121 18:31:29.125853 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba1dbe7c160275afae696b07a62f00b7eb61a955782adf02caefbd95939b1493\": container with ID starting with ba1dbe7c160275afae696b07a62f00b7eb61a955782adf02caefbd95939b1493 not found: ID does not exist" containerID="ba1dbe7c160275afae696b07a62f00b7eb61a955782adf02caefbd95939b1493" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.125889 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba1dbe7c160275afae696b07a62f00b7eb61a955782adf02caefbd95939b1493"} err="failed to get container status \"ba1dbe7c160275afae696b07a62f00b7eb61a955782adf02caefbd95939b1493\": rpc error: code = NotFound desc = could not find container \"ba1dbe7c160275afae696b07a62f00b7eb61a955782adf02caefbd95939b1493\": container with ID starting with ba1dbe7c160275afae696b07a62f00b7eb61a955782adf02caefbd95939b1493 not found: ID does not exist" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.125913 4808 scope.go:117] "RemoveContainer" containerID="6624bc1e34644ec644da76e9ba8e3fedf50bd4481db75a8fc10f7b987cf03d4a" Jan 21 18:31:29 crc kubenswrapper[4808]: E0121 18:31:29.126356 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6624bc1e34644ec644da76e9ba8e3fedf50bd4481db75a8fc10f7b987cf03d4a\": container with ID starting with 6624bc1e34644ec644da76e9ba8e3fedf50bd4481db75a8fc10f7b987cf03d4a not found: ID does not exist" containerID="6624bc1e34644ec644da76e9ba8e3fedf50bd4481db75a8fc10f7b987cf03d4a" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.126403 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6624bc1e34644ec644da76e9ba8e3fedf50bd4481db75a8fc10f7b987cf03d4a"} err="failed to get container status \"6624bc1e34644ec644da76e9ba8e3fedf50bd4481db75a8fc10f7b987cf03d4a\": rpc error: code = NotFound desc = could not find container \"6624bc1e34644ec644da76e9ba8e3fedf50bd4481db75a8fc10f7b987cf03d4a\": container with ID starting with 6624bc1e34644ec644da76e9ba8e3fedf50bd4481db75a8fc10f7b987cf03d4a not found: ID does not exist" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.126432 4808 scope.go:117] "RemoveContainer" containerID="1c2540cfd9f6f2aa5e9c8226454898bf6f7394cd1609b32776dbc9ddc106fc74" Jan 21 18:31:29 crc kubenswrapper[4808]: E0121 18:31:29.126958 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c2540cfd9f6f2aa5e9c8226454898bf6f7394cd1609b32776dbc9ddc106fc74\": container with ID starting with 1c2540cfd9f6f2aa5e9c8226454898bf6f7394cd1609b32776dbc9ddc106fc74 not found: ID does not exist" containerID="1c2540cfd9f6f2aa5e9c8226454898bf6f7394cd1609b32776dbc9ddc106fc74" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.126983 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c2540cfd9f6f2aa5e9c8226454898bf6f7394cd1609b32776dbc9ddc106fc74"} err="failed to get container status \"1c2540cfd9f6f2aa5e9c8226454898bf6f7394cd1609b32776dbc9ddc106fc74\": rpc error: code = NotFound desc = could not find container \"1c2540cfd9f6f2aa5e9c8226454898bf6f7394cd1609b32776dbc9ddc106fc74\": container with ID starting with 1c2540cfd9f6f2aa5e9c8226454898bf6f7394cd1609b32776dbc9ddc106fc74 not found: ID does not exist" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.127000 4808 scope.go:117] "RemoveContainer" containerID="fa7c5bf063b9350cd7c662bbd4575389c72037d13e2efddf22bc2e186973b0cb" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.128668 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cbwpb"] Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.135875 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cbwpb"] Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.144154 4808 scope.go:117] "RemoveContainer" containerID="3845f363524ae982813cf7f417027b238906b51381442ab15ce25a9390788ab0" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.165272 4808 scope.go:117] "RemoveContainer" containerID="4d2be47b39e77a53c6b82949dfa1368760aa914be579e7af148f4593039c2521" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.186902 4808 scope.go:117] "RemoveContainer" containerID="fa7c5bf063b9350cd7c662bbd4575389c72037d13e2efddf22bc2e186973b0cb" Jan 21 18:31:29 crc kubenswrapper[4808]: E0121 18:31:29.188434 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa7c5bf063b9350cd7c662bbd4575389c72037d13e2efddf22bc2e186973b0cb\": container with ID starting with fa7c5bf063b9350cd7c662bbd4575389c72037d13e2efddf22bc2e186973b0cb not found: ID does not exist" containerID="fa7c5bf063b9350cd7c662bbd4575389c72037d13e2efddf22bc2e186973b0cb" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.188492 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa7c5bf063b9350cd7c662bbd4575389c72037d13e2efddf22bc2e186973b0cb"} err="failed to get container status \"fa7c5bf063b9350cd7c662bbd4575389c72037d13e2efddf22bc2e186973b0cb\": rpc error: code = NotFound desc = could not find container \"fa7c5bf063b9350cd7c662bbd4575389c72037d13e2efddf22bc2e186973b0cb\": container with ID starting with fa7c5bf063b9350cd7c662bbd4575389c72037d13e2efddf22bc2e186973b0cb not found: ID does not exist" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.188524 4808 scope.go:117] "RemoveContainer" containerID="3845f363524ae982813cf7f417027b238906b51381442ab15ce25a9390788ab0" Jan 21 18:31:29 crc kubenswrapper[4808]: E0121 18:31:29.188890 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3845f363524ae982813cf7f417027b238906b51381442ab15ce25a9390788ab0\": container with ID starting with 3845f363524ae982813cf7f417027b238906b51381442ab15ce25a9390788ab0 not found: ID does not exist" containerID="3845f363524ae982813cf7f417027b238906b51381442ab15ce25a9390788ab0" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.188930 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3845f363524ae982813cf7f417027b238906b51381442ab15ce25a9390788ab0"} err="failed to get container status \"3845f363524ae982813cf7f417027b238906b51381442ab15ce25a9390788ab0\": rpc error: code = NotFound desc = could not find container \"3845f363524ae982813cf7f417027b238906b51381442ab15ce25a9390788ab0\": container with ID starting with 3845f363524ae982813cf7f417027b238906b51381442ab15ce25a9390788ab0 not found: ID does not exist" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.188953 4808 scope.go:117] "RemoveContainer" containerID="4d2be47b39e77a53c6b82949dfa1368760aa914be579e7af148f4593039c2521" Jan 21 18:31:29 crc kubenswrapper[4808]: E0121 18:31:29.189199 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d2be47b39e77a53c6b82949dfa1368760aa914be579e7af148f4593039c2521\": container with ID starting with 4d2be47b39e77a53c6b82949dfa1368760aa914be579e7af148f4593039c2521 not found: ID does not exist" containerID="4d2be47b39e77a53c6b82949dfa1368760aa914be579e7af148f4593039c2521" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.189219 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d2be47b39e77a53c6b82949dfa1368760aa914be579e7af148f4593039c2521"} err="failed to get container status \"4d2be47b39e77a53c6b82949dfa1368760aa914be579e7af148f4593039c2521\": rpc error: code = NotFound desc = could not find container \"4d2be47b39e77a53c6b82949dfa1368760aa914be579e7af148f4593039c2521\": container with ID starting with 4d2be47b39e77a53c6b82949dfa1368760aa914be579e7af148f4593039c2521 not found: ID does not exist" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.189235 4808 scope.go:117] "RemoveContainer" containerID="7d1abbc03aa42016ed07b245e1a1ec9e616b4f93dca34fc1da743e210928b23b" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.209840 4808 scope.go:117] "RemoveContainer" containerID="08eadee2b87a03f92f78f5ddf1f293085e3cbb7401c20a8b01b423021c463bac" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.223792 4808 scope.go:117] "RemoveContainer" containerID="099e0d2b82bdcb9221be0dd433b35617d145bbc57a883d09664d085ded5f25a5" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.242272 4808 scope.go:117] "RemoveContainer" containerID="7d1abbc03aa42016ed07b245e1a1ec9e616b4f93dca34fc1da743e210928b23b" Jan 21 18:31:29 crc kubenswrapper[4808]: E0121 18:31:29.242757 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d1abbc03aa42016ed07b245e1a1ec9e616b4f93dca34fc1da743e210928b23b\": container with ID starting with 7d1abbc03aa42016ed07b245e1a1ec9e616b4f93dca34fc1da743e210928b23b not found: ID does not exist" containerID="7d1abbc03aa42016ed07b245e1a1ec9e616b4f93dca34fc1da743e210928b23b" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.242800 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d1abbc03aa42016ed07b245e1a1ec9e616b4f93dca34fc1da743e210928b23b"} err="failed to get container status \"7d1abbc03aa42016ed07b245e1a1ec9e616b4f93dca34fc1da743e210928b23b\": rpc error: code = NotFound desc = could not find container \"7d1abbc03aa42016ed07b245e1a1ec9e616b4f93dca34fc1da743e210928b23b\": container with ID starting with 7d1abbc03aa42016ed07b245e1a1ec9e616b4f93dca34fc1da743e210928b23b not found: ID does not exist" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.242829 4808 scope.go:117] "RemoveContainer" containerID="08eadee2b87a03f92f78f5ddf1f293085e3cbb7401c20a8b01b423021c463bac" Jan 21 18:31:29 crc kubenswrapper[4808]: E0121 18:31:29.243157 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08eadee2b87a03f92f78f5ddf1f293085e3cbb7401c20a8b01b423021c463bac\": container with ID starting with 08eadee2b87a03f92f78f5ddf1f293085e3cbb7401c20a8b01b423021c463bac not found: ID does not exist" containerID="08eadee2b87a03f92f78f5ddf1f293085e3cbb7401c20a8b01b423021c463bac" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.243194 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08eadee2b87a03f92f78f5ddf1f293085e3cbb7401c20a8b01b423021c463bac"} err="failed to get container status \"08eadee2b87a03f92f78f5ddf1f293085e3cbb7401c20a8b01b423021c463bac\": rpc error: code = NotFound desc = could not find container \"08eadee2b87a03f92f78f5ddf1f293085e3cbb7401c20a8b01b423021c463bac\": container with ID starting with 08eadee2b87a03f92f78f5ddf1f293085e3cbb7401c20a8b01b423021c463bac not found: ID does not exist" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.243236 4808 scope.go:117] "RemoveContainer" containerID="099e0d2b82bdcb9221be0dd433b35617d145bbc57a883d09664d085ded5f25a5" Jan 21 18:31:29 crc kubenswrapper[4808]: E0121 18:31:29.243568 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"099e0d2b82bdcb9221be0dd433b35617d145bbc57a883d09664d085ded5f25a5\": container with ID starting with 099e0d2b82bdcb9221be0dd433b35617d145bbc57a883d09664d085ded5f25a5 not found: ID does not exist" containerID="099e0d2b82bdcb9221be0dd433b35617d145bbc57a883d09664d085ded5f25a5" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.243594 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"099e0d2b82bdcb9221be0dd433b35617d145bbc57a883d09664d085ded5f25a5"} err="failed to get container status \"099e0d2b82bdcb9221be0dd433b35617d145bbc57a883d09664d085ded5f25a5\": rpc error: code = NotFound desc = could not find container \"099e0d2b82bdcb9221be0dd433b35617d145bbc57a883d09664d085ded5f25a5\": container with ID starting with 099e0d2b82bdcb9221be0dd433b35617d145bbc57a883d09664d085ded5f25a5 not found: ID does not exist" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.243608 4808 scope.go:117] "RemoveContainer" containerID="48bce33ad57d52f0f446058cfd1ba8b2673b597e34eed6ca33bb88dda0af606c" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.258871 4808 scope.go:117] "RemoveContainer" containerID="515f64447fd320d6815e2fc89ec6992b1efcd2fed75bcc9bab0e96bce7064366" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.276340 4808 scope.go:117] "RemoveContainer" containerID="aa5669676685ea416374092bb35b3e74e779c7133d7b0a8a58c00ce8fbe57f0b" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.290465 4808 scope.go:117] "RemoveContainer" containerID="48bce33ad57d52f0f446058cfd1ba8b2673b597e34eed6ca33bb88dda0af606c" Jan 21 18:31:29 crc kubenswrapper[4808]: E0121 18:31:29.290875 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48bce33ad57d52f0f446058cfd1ba8b2673b597e34eed6ca33bb88dda0af606c\": container with ID starting with 48bce33ad57d52f0f446058cfd1ba8b2673b597e34eed6ca33bb88dda0af606c not found: ID does not exist" containerID="48bce33ad57d52f0f446058cfd1ba8b2673b597e34eed6ca33bb88dda0af606c" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.290918 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48bce33ad57d52f0f446058cfd1ba8b2673b597e34eed6ca33bb88dda0af606c"} err="failed to get container status \"48bce33ad57d52f0f446058cfd1ba8b2673b597e34eed6ca33bb88dda0af606c\": rpc error: code = NotFound desc = could not find container \"48bce33ad57d52f0f446058cfd1ba8b2673b597e34eed6ca33bb88dda0af606c\": container with ID starting with 48bce33ad57d52f0f446058cfd1ba8b2673b597e34eed6ca33bb88dda0af606c not found: ID does not exist" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.290948 4808 scope.go:117] "RemoveContainer" containerID="515f64447fd320d6815e2fc89ec6992b1efcd2fed75bcc9bab0e96bce7064366" Jan 21 18:31:29 crc kubenswrapper[4808]: E0121 18:31:29.291321 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"515f64447fd320d6815e2fc89ec6992b1efcd2fed75bcc9bab0e96bce7064366\": container with ID starting with 515f64447fd320d6815e2fc89ec6992b1efcd2fed75bcc9bab0e96bce7064366 not found: ID does not exist" containerID="515f64447fd320d6815e2fc89ec6992b1efcd2fed75bcc9bab0e96bce7064366" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.291350 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"515f64447fd320d6815e2fc89ec6992b1efcd2fed75bcc9bab0e96bce7064366"} err="failed to get container status \"515f64447fd320d6815e2fc89ec6992b1efcd2fed75bcc9bab0e96bce7064366\": rpc error: code = NotFound desc = could not find container \"515f64447fd320d6815e2fc89ec6992b1efcd2fed75bcc9bab0e96bce7064366\": container with ID starting with 515f64447fd320d6815e2fc89ec6992b1efcd2fed75bcc9bab0e96bce7064366 not found: ID does not exist" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.291373 4808 scope.go:117] "RemoveContainer" containerID="aa5669676685ea416374092bb35b3e74e779c7133d7b0a8a58c00ce8fbe57f0b" Jan 21 18:31:29 crc kubenswrapper[4808]: E0121 18:31:29.291585 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa5669676685ea416374092bb35b3e74e779c7133d7b0a8a58c00ce8fbe57f0b\": container with ID starting with aa5669676685ea416374092bb35b3e74e779c7133d7b0a8a58c00ce8fbe57f0b not found: ID does not exist" containerID="aa5669676685ea416374092bb35b3e74e779c7133d7b0a8a58c00ce8fbe57f0b" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.291610 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa5669676685ea416374092bb35b3e74e779c7133d7b0a8a58c00ce8fbe57f0b"} err="failed to get container status \"aa5669676685ea416374092bb35b3e74e779c7133d7b0a8a58c00ce8fbe57f0b\": rpc error: code = NotFound desc = could not find container \"aa5669676685ea416374092bb35b3e74e779c7133d7b0a8a58c00ce8fbe57f0b\": container with ID starting with aa5669676685ea416374092bb35b3e74e779c7133d7b0a8a58c00ce8fbe57f0b not found: ID does not exist" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.291624 4808 scope.go:117] "RemoveContainer" containerID="adf91b787ad2ac75272151d18611c4d9cbaf09e66530001a605d177fa709ff7c" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.306041 4808 scope.go:117] "RemoveContainer" containerID="d43fb31c253ba16440f65cf1b28595d677bbe5ce579b6a51afc16f3ec4218d04" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.323153 4808 scope.go:117] "RemoveContainer" containerID="a49b3340884343001efa7248d836b1be90456f2a185cc7ddc826cc0ce8b632b3" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.339164 4808 scope.go:117] "RemoveContainer" containerID="adf91b787ad2ac75272151d18611c4d9cbaf09e66530001a605d177fa709ff7c" Jan 21 18:31:29 crc kubenswrapper[4808]: E0121 18:31:29.339580 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adf91b787ad2ac75272151d18611c4d9cbaf09e66530001a605d177fa709ff7c\": container with ID starting with adf91b787ad2ac75272151d18611c4d9cbaf09e66530001a605d177fa709ff7c not found: ID does not exist" containerID="adf91b787ad2ac75272151d18611c4d9cbaf09e66530001a605d177fa709ff7c" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.339627 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adf91b787ad2ac75272151d18611c4d9cbaf09e66530001a605d177fa709ff7c"} err="failed to get container status \"adf91b787ad2ac75272151d18611c4d9cbaf09e66530001a605d177fa709ff7c\": rpc error: code = NotFound desc = could not find container \"adf91b787ad2ac75272151d18611c4d9cbaf09e66530001a605d177fa709ff7c\": container with ID starting with adf91b787ad2ac75272151d18611c4d9cbaf09e66530001a605d177fa709ff7c not found: ID does not exist" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.339658 4808 scope.go:117] "RemoveContainer" containerID="d43fb31c253ba16440f65cf1b28595d677bbe5ce579b6a51afc16f3ec4218d04" Jan 21 18:31:29 crc kubenswrapper[4808]: E0121 18:31:29.340054 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d43fb31c253ba16440f65cf1b28595d677bbe5ce579b6a51afc16f3ec4218d04\": container with ID starting with d43fb31c253ba16440f65cf1b28595d677bbe5ce579b6a51afc16f3ec4218d04 not found: ID does not exist" containerID="d43fb31c253ba16440f65cf1b28595d677bbe5ce579b6a51afc16f3ec4218d04" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.340110 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d43fb31c253ba16440f65cf1b28595d677bbe5ce579b6a51afc16f3ec4218d04"} err="failed to get container status \"d43fb31c253ba16440f65cf1b28595d677bbe5ce579b6a51afc16f3ec4218d04\": rpc error: code = NotFound desc = could not find container \"d43fb31c253ba16440f65cf1b28595d677bbe5ce579b6a51afc16f3ec4218d04\": container with ID starting with d43fb31c253ba16440f65cf1b28595d677bbe5ce579b6a51afc16f3ec4218d04 not found: ID does not exist" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.340152 4808 scope.go:117] "RemoveContainer" containerID="a49b3340884343001efa7248d836b1be90456f2a185cc7ddc826cc0ce8b632b3" Jan 21 18:31:29 crc kubenswrapper[4808]: E0121 18:31:29.340628 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a49b3340884343001efa7248d836b1be90456f2a185cc7ddc826cc0ce8b632b3\": container with ID starting with a49b3340884343001efa7248d836b1be90456f2a185cc7ddc826cc0ce8b632b3 not found: ID does not exist" containerID="a49b3340884343001efa7248d836b1be90456f2a185cc7ddc826cc0ce8b632b3" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.340656 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a49b3340884343001efa7248d836b1be90456f2a185cc7ddc826cc0ce8b632b3"} err="failed to get container status \"a49b3340884343001efa7248d836b1be90456f2a185cc7ddc826cc0ce8b632b3\": rpc error: code = NotFound desc = could not find container \"a49b3340884343001efa7248d836b1be90456f2a185cc7ddc826cc0ce8b632b3\": container with ID starting with a49b3340884343001efa7248d836b1be90456f2a185cc7ddc826cc0ce8b632b3 not found: ID does not exist" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.527956 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="173b6b54-1024-4806-8056-985010126525" path="/var/lib/kubelet/pods/173b6b54-1024-4806-8056-985010126525/volumes" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.528626 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1df03daa-d900-4a6d-9776-26f3b05843ce" path="/var/lib/kubelet/pods/1df03daa-d900-4a6d-9776-26f3b05843ce/volumes" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.529191 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21ebff69-0bfe-427d-8330-4e454159dd01" path="/var/lib/kubelet/pods/21ebff69-0bfe-427d-8330-4e454159dd01/volumes" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.530196 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62cdb877-c15b-4990-8820-3dcc6ffc0b04" path="/var/lib/kubelet/pods/62cdb877-c15b-4990-8820-3dcc6ffc0b04/volumes" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.530801 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f554b03-f857-4309-ba4c-91e8c607a14d" path="/var/lib/kubelet/pods/7f554b03-f857-4309-ba4c-91e8c607a14d/volumes" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.531833 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d1c03b0-e10f-47ae-841c-153e4ae0f9cd" path="/var/lib/kubelet/pods/9d1c03b0-e10f-47ae-841c-153e4ae0f9cd/volumes" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.532404 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5322a4c-c457-48c5-9c8d-6ce242b4b87b" path="/var/lib/kubelet/pods/c5322a4c-c457-48c5-9c8d-6ce242b4b87b/volumes" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.532940 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dde2d0c2-70a4-4d0a-bc7a-29fd72c50095" path="/var/lib/kubelet/pods/dde2d0c2-70a4-4d0a-bc7a-29fd72c50095/volumes" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.601130 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.601498 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.601549 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.602211 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dd13afd7fd43a6e8f7732443fd57a881906abc4ac14c52babeb4773a1dc24aa8"} pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.602307 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" containerID="cri-o://dd13afd7fd43a6e8f7732443fd57a881906abc4ac14c52babeb4773a1dc24aa8" gracePeriod=600 Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.831142 4808 generic.go:334] "Generic (PLEG): container finished" podID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerID="dd13afd7fd43a6e8f7732443fd57a881906abc4ac14c52babeb4773a1dc24aa8" exitCode=0 Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.831230 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerDied","Data":"dd13afd7fd43a6e8f7732443fd57a881906abc4ac14c52babeb4773a1dc24aa8"} Jan 21 18:31:29 crc kubenswrapper[4808]: I0121 18:31:29.831315 4808 scope.go:117] "RemoveContainer" containerID="3117a4da6dd06f7fb3ce5cdb42bbd4fa0336fafde5cee9ce9b3732ec5d2f42f1" Jan 21 18:31:30 crc kubenswrapper[4808]: I0121 18:31:30.868214 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerStarted","Data":"f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0"} Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.426214 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7hrw6"] Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427258 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d1c03b0-e10f-47ae-841c-153e4ae0f9cd" containerName="extract-content" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427273 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d1c03b0-e10f-47ae-841c-153e4ae0f9cd" containerName="extract-content" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427284 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dde2d0c2-70a4-4d0a-bc7a-29fd72c50095" containerName="extract-content" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427290 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="dde2d0c2-70a4-4d0a-bc7a-29fd72c50095" containerName="extract-content" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427305 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="173b6b54-1024-4806-8056-985010126525" containerName="extract-utilities" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427311 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="173b6b54-1024-4806-8056-985010126525" containerName="extract-utilities" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427319 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dde2d0c2-70a4-4d0a-bc7a-29fd72c50095" containerName="extract-utilities" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427325 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="dde2d0c2-70a4-4d0a-bc7a-29fd72c50095" containerName="extract-utilities" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427334 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21ebff69-0bfe-427d-8330-4e454159dd01" containerName="extract-content" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427340 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="21ebff69-0bfe-427d-8330-4e454159dd01" containerName="extract-content" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427346 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5322a4c-c457-48c5-9c8d-6ce242b4b87b" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427351 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5322a4c-c457-48c5-9c8d-6ce242b4b87b" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427359 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="173b6b54-1024-4806-8056-985010126525" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427365 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="173b6b54-1024-4806-8056-985010126525" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427374 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df03daa-d900-4a6d-9776-26f3b05843ce" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427379 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df03daa-d900-4a6d-9776-26f3b05843ce" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427388 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d1c03b0-e10f-47ae-841c-153e4ae0f9cd" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427394 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d1c03b0-e10f-47ae-841c-153e4ae0f9cd" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427402 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62cdb877-c15b-4990-8820-3dcc6ffc0b04" containerName="extract-utilities" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427409 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="62cdb877-c15b-4990-8820-3dcc6ffc0b04" containerName="extract-utilities" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427419 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df03daa-d900-4a6d-9776-26f3b05843ce" containerName="extract-content" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427424 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df03daa-d900-4a6d-9776-26f3b05843ce" containerName="extract-content" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427431 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21ebff69-0bfe-427d-8330-4e454159dd01" containerName="extract-utilities" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427437 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="21ebff69-0bfe-427d-8330-4e454159dd01" containerName="extract-utilities" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427445 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f554b03-f857-4309-ba4c-91e8c607a14d" containerName="extract-content" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427451 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f554b03-f857-4309-ba4c-91e8c607a14d" containerName="extract-content" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427460 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5322a4c-c457-48c5-9c8d-6ce242b4b87b" containerName="extract-utilities" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427465 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5322a4c-c457-48c5-9c8d-6ce242b4b87b" containerName="extract-utilities" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427475 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62cdb877-c15b-4990-8820-3dcc6ffc0b04" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427481 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="62cdb877-c15b-4990-8820-3dcc6ffc0b04" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427487 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df03daa-d900-4a6d-9776-26f3b05843ce" containerName="extract-utilities" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427493 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df03daa-d900-4a6d-9776-26f3b05843ce" containerName="extract-utilities" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427501 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f554b03-f857-4309-ba4c-91e8c607a14d" containerName="extract-utilities" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427507 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f554b03-f857-4309-ba4c-91e8c607a14d" containerName="extract-utilities" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427514 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5322a4c-c457-48c5-9c8d-6ce242b4b87b" containerName="extract-content" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427521 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5322a4c-c457-48c5-9c8d-6ce242b4b87b" containerName="extract-content" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427529 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62cdb877-c15b-4990-8820-3dcc6ffc0b04" containerName="extract-content" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427535 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="62cdb877-c15b-4990-8820-3dcc6ffc0b04" containerName="extract-content" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427545 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21ebff69-0bfe-427d-8330-4e454159dd01" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427551 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="21ebff69-0bfe-427d-8330-4e454159dd01" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427561 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f554b03-f857-4309-ba4c-91e8c607a14d" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427566 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f554b03-f857-4309-ba4c-91e8c607a14d" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427573 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="173b6b54-1024-4806-8056-985010126525" containerName="extract-content" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427579 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="173b6b54-1024-4806-8056-985010126525" containerName="extract-content" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427587 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dde2d0c2-70a4-4d0a-bc7a-29fd72c50095" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427593 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="dde2d0c2-70a4-4d0a-bc7a-29fd72c50095" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: E0121 18:31:49.427603 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d1c03b0-e10f-47ae-841c-153e4ae0f9cd" containerName="extract-utilities" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427609 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d1c03b0-e10f-47ae-841c-153e4ae0f9cd" containerName="extract-utilities" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427706 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="1df03daa-d900-4a6d-9776-26f3b05843ce" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427719 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="62cdb877-c15b-4990-8820-3dcc6ffc0b04" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427729 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="173b6b54-1024-4806-8056-985010126525" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427742 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="21ebff69-0bfe-427d-8330-4e454159dd01" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427757 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="dde2d0c2-70a4-4d0a-bc7a-29fd72c50095" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427768 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5322a4c-c457-48c5-9c8d-6ce242b4b87b" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427777 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d1c03b0-e10f-47ae-841c-153e4ae0f9cd" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.427784 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f554b03-f857-4309-ba4c-91e8c607a14d" containerName="registry-server" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.429576 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7hrw6" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.440815 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7hrw6"] Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.530027 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2157473d-7ad3-45a3-8215-dbf88187ae6c-catalog-content\") pod \"community-operators-7hrw6\" (UID: \"2157473d-7ad3-45a3-8215-dbf88187ae6c\") " pod="openshift-marketplace/community-operators-7hrw6" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.530077 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f72h6\" (UniqueName: \"kubernetes.io/projected/2157473d-7ad3-45a3-8215-dbf88187ae6c-kube-api-access-f72h6\") pod \"community-operators-7hrw6\" (UID: \"2157473d-7ad3-45a3-8215-dbf88187ae6c\") " pod="openshift-marketplace/community-operators-7hrw6" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.530133 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2157473d-7ad3-45a3-8215-dbf88187ae6c-utilities\") pod \"community-operators-7hrw6\" (UID: \"2157473d-7ad3-45a3-8215-dbf88187ae6c\") " pod="openshift-marketplace/community-operators-7hrw6" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.631689 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2157473d-7ad3-45a3-8215-dbf88187ae6c-catalog-content\") pod \"community-operators-7hrw6\" (UID: \"2157473d-7ad3-45a3-8215-dbf88187ae6c\") " pod="openshift-marketplace/community-operators-7hrw6" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.631744 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f72h6\" (UniqueName: \"kubernetes.io/projected/2157473d-7ad3-45a3-8215-dbf88187ae6c-kube-api-access-f72h6\") pod \"community-operators-7hrw6\" (UID: \"2157473d-7ad3-45a3-8215-dbf88187ae6c\") " pod="openshift-marketplace/community-operators-7hrw6" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.631808 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2157473d-7ad3-45a3-8215-dbf88187ae6c-utilities\") pod \"community-operators-7hrw6\" (UID: \"2157473d-7ad3-45a3-8215-dbf88187ae6c\") " pod="openshift-marketplace/community-operators-7hrw6" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.632290 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2157473d-7ad3-45a3-8215-dbf88187ae6c-catalog-content\") pod \"community-operators-7hrw6\" (UID: \"2157473d-7ad3-45a3-8215-dbf88187ae6c\") " pod="openshift-marketplace/community-operators-7hrw6" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.632363 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2157473d-7ad3-45a3-8215-dbf88187ae6c-utilities\") pod \"community-operators-7hrw6\" (UID: \"2157473d-7ad3-45a3-8215-dbf88187ae6c\") " pod="openshift-marketplace/community-operators-7hrw6" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.653810 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f72h6\" (UniqueName: \"kubernetes.io/projected/2157473d-7ad3-45a3-8215-dbf88187ae6c-kube-api-access-f72h6\") pod \"community-operators-7hrw6\" (UID: \"2157473d-7ad3-45a3-8215-dbf88187ae6c\") " pod="openshift-marketplace/community-operators-7hrw6" Jan 21 18:31:49 crc kubenswrapper[4808]: I0121 18:31:49.788181 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7hrw6" Jan 21 18:31:50 crc kubenswrapper[4808]: I0121 18:31:50.304102 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7hrw6"] Jan 21 18:31:51 crc kubenswrapper[4808]: I0121 18:31:51.014104 4808 generic.go:334] "Generic (PLEG): container finished" podID="2157473d-7ad3-45a3-8215-dbf88187ae6c" containerID="f662727a1fe5f11908a3627b1827ee843587083abb4bad7930ea9ac8f257ecfb" exitCode=0 Jan 21 18:31:51 crc kubenswrapper[4808]: I0121 18:31:51.014147 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hrw6" event={"ID":"2157473d-7ad3-45a3-8215-dbf88187ae6c","Type":"ContainerDied","Data":"f662727a1fe5f11908a3627b1827ee843587083abb4bad7930ea9ac8f257ecfb"} Jan 21 18:31:51 crc kubenswrapper[4808]: I0121 18:31:51.014404 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hrw6" event={"ID":"2157473d-7ad3-45a3-8215-dbf88187ae6c","Type":"ContainerStarted","Data":"f01aa2f04c53779abef371f3ffe6307d68c25421ea7c68b6bc51a39e1d50340b"} Jan 21 18:31:53 crc kubenswrapper[4808]: I0121 18:31:53.031628 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hrw6" event={"ID":"2157473d-7ad3-45a3-8215-dbf88187ae6c","Type":"ContainerStarted","Data":"027c60e4f90074860187e100e9915c3d34d362d276264049bf13c84062869f42"} Jan 21 18:31:54 crc kubenswrapper[4808]: I0121 18:31:54.042016 4808 generic.go:334] "Generic (PLEG): container finished" podID="2157473d-7ad3-45a3-8215-dbf88187ae6c" containerID="027c60e4f90074860187e100e9915c3d34d362d276264049bf13c84062869f42" exitCode=0 Jan 21 18:31:54 crc kubenswrapper[4808]: I0121 18:31:54.042114 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hrw6" event={"ID":"2157473d-7ad3-45a3-8215-dbf88187ae6c","Type":"ContainerDied","Data":"027c60e4f90074860187e100e9915c3d34d362d276264049bf13c84062869f42"} Jan 21 18:31:55 crc kubenswrapper[4808]: I0121 18:31:55.053962 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hrw6" event={"ID":"2157473d-7ad3-45a3-8215-dbf88187ae6c","Type":"ContainerStarted","Data":"c3af51d207d215ec7b88f821fd62556b6a7e67b7a0f7e9ae95d9caac4396b3ee"} Jan 21 18:31:55 crc kubenswrapper[4808]: I0121 18:31:55.086741 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7hrw6" podStartSLOduration=3.544799314 podStartE2EDuration="6.086717378s" podCreationTimestamp="2026-01-21 18:31:49 +0000 UTC" firstStartedPulling="2026-01-21 18:31:52.023541985 +0000 UTC m=+2146.754052880" lastFinishedPulling="2026-01-21 18:31:54.565460059 +0000 UTC m=+2149.295970944" observedRunningTime="2026-01-21 18:31:55.081334846 +0000 UTC m=+2149.811845731" watchObservedRunningTime="2026-01-21 18:31:55.086717378 +0000 UTC m=+2149.817228263" Jan 21 18:31:59 crc kubenswrapper[4808]: I0121 18:31:59.788481 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7hrw6" Jan 21 18:31:59 crc kubenswrapper[4808]: I0121 18:31:59.789391 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7hrw6" Jan 21 18:31:59 crc kubenswrapper[4808]: I0121 18:31:59.831916 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7hrw6" Jan 21 18:32:00 crc kubenswrapper[4808]: I0121 18:32:00.128353 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7hrw6" Jan 21 18:32:00 crc kubenswrapper[4808]: I0121 18:32:00.177916 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7hrw6"] Jan 21 18:32:02 crc kubenswrapper[4808]: I0121 18:32:02.100500 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7hrw6" podUID="2157473d-7ad3-45a3-8215-dbf88187ae6c" containerName="registry-server" containerID="cri-o://c3af51d207d215ec7b88f821fd62556b6a7e67b7a0f7e9ae95d9caac4396b3ee" gracePeriod=2 Jan 21 18:32:04 crc kubenswrapper[4808]: I0121 18:32:04.132400 4808 generic.go:334] "Generic (PLEG): container finished" podID="2157473d-7ad3-45a3-8215-dbf88187ae6c" containerID="c3af51d207d215ec7b88f821fd62556b6a7e67b7a0f7e9ae95d9caac4396b3ee" exitCode=0 Jan 21 18:32:04 crc kubenswrapper[4808]: I0121 18:32:04.132483 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hrw6" event={"ID":"2157473d-7ad3-45a3-8215-dbf88187ae6c","Type":"ContainerDied","Data":"c3af51d207d215ec7b88f821fd62556b6a7e67b7a0f7e9ae95d9caac4396b3ee"} Jan 21 18:32:05 crc kubenswrapper[4808]: I0121 18:32:05.142789 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hrw6" event={"ID":"2157473d-7ad3-45a3-8215-dbf88187ae6c","Type":"ContainerDied","Data":"f01aa2f04c53779abef371f3ffe6307d68c25421ea7c68b6bc51a39e1d50340b"} Jan 21 18:32:05 crc kubenswrapper[4808]: I0121 18:32:05.143165 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f01aa2f04c53779abef371f3ffe6307d68c25421ea7c68b6bc51a39e1d50340b" Jan 21 18:32:05 crc kubenswrapper[4808]: I0121 18:32:05.144409 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7hrw6" Jan 21 18:32:05 crc kubenswrapper[4808]: I0121 18:32:05.255736 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2157473d-7ad3-45a3-8215-dbf88187ae6c-catalog-content\") pod \"2157473d-7ad3-45a3-8215-dbf88187ae6c\" (UID: \"2157473d-7ad3-45a3-8215-dbf88187ae6c\") " Jan 21 18:32:05 crc kubenswrapper[4808]: I0121 18:32:05.255776 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f72h6\" (UniqueName: \"kubernetes.io/projected/2157473d-7ad3-45a3-8215-dbf88187ae6c-kube-api-access-f72h6\") pod \"2157473d-7ad3-45a3-8215-dbf88187ae6c\" (UID: \"2157473d-7ad3-45a3-8215-dbf88187ae6c\") " Jan 21 18:32:05 crc kubenswrapper[4808]: I0121 18:32:05.255871 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2157473d-7ad3-45a3-8215-dbf88187ae6c-utilities\") pod \"2157473d-7ad3-45a3-8215-dbf88187ae6c\" (UID: \"2157473d-7ad3-45a3-8215-dbf88187ae6c\") " Jan 21 18:32:05 crc kubenswrapper[4808]: I0121 18:32:05.262773 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2157473d-7ad3-45a3-8215-dbf88187ae6c-kube-api-access-f72h6" (OuterVolumeSpecName: "kube-api-access-f72h6") pod "2157473d-7ad3-45a3-8215-dbf88187ae6c" (UID: "2157473d-7ad3-45a3-8215-dbf88187ae6c"). InnerVolumeSpecName "kube-api-access-f72h6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:32:05 crc kubenswrapper[4808]: I0121 18:32:05.263732 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2157473d-7ad3-45a3-8215-dbf88187ae6c-utilities" (OuterVolumeSpecName: "utilities") pod "2157473d-7ad3-45a3-8215-dbf88187ae6c" (UID: "2157473d-7ad3-45a3-8215-dbf88187ae6c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:32:05 crc kubenswrapper[4808]: I0121 18:32:05.307828 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2157473d-7ad3-45a3-8215-dbf88187ae6c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2157473d-7ad3-45a3-8215-dbf88187ae6c" (UID: "2157473d-7ad3-45a3-8215-dbf88187ae6c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:32:05 crc kubenswrapper[4808]: I0121 18:32:05.356995 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2157473d-7ad3-45a3-8215-dbf88187ae6c-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:32:05 crc kubenswrapper[4808]: I0121 18:32:05.357027 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2157473d-7ad3-45a3-8215-dbf88187ae6c-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:32:05 crc kubenswrapper[4808]: I0121 18:32:05.357039 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f72h6\" (UniqueName: \"kubernetes.io/projected/2157473d-7ad3-45a3-8215-dbf88187ae6c-kube-api-access-f72h6\") on node \"crc\" DevicePath \"\"" Jan 21 18:32:06 crc kubenswrapper[4808]: I0121 18:32:06.150112 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7hrw6" Jan 21 18:32:06 crc kubenswrapper[4808]: I0121 18:32:06.176521 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7hrw6"] Jan 21 18:32:06 crc kubenswrapper[4808]: I0121 18:32:06.183567 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7hrw6"] Jan 21 18:32:07 crc kubenswrapper[4808]: I0121 18:32:07.534304 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2157473d-7ad3-45a3-8215-dbf88187ae6c" path="/var/lib/kubelet/pods/2157473d-7ad3-45a3-8215-dbf88187ae6c/volumes" Jan 21 18:33:29 crc kubenswrapper[4808]: I0121 18:33:29.600108 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:33:29 crc kubenswrapper[4808]: I0121 18:33:29.600697 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:33:59 crc kubenswrapper[4808]: I0121 18:33:59.599135 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:33:59 crc kubenswrapper[4808]: I0121 18:33:59.599747 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:34:29 crc kubenswrapper[4808]: I0121 18:34:29.600093 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:34:29 crc kubenswrapper[4808]: I0121 18:34:29.600817 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:34:29 crc kubenswrapper[4808]: I0121 18:34:29.600877 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 18:34:29 crc kubenswrapper[4808]: I0121 18:34:29.601684 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0"} pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 18:34:29 crc kubenswrapper[4808]: I0121 18:34:29.601785 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" containerID="cri-o://f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" gracePeriod=600 Jan 21 18:34:29 crc kubenswrapper[4808]: E0121 18:34:29.729052 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:34:30 crc kubenswrapper[4808]: I0121 18:34:30.158476 4808 generic.go:334] "Generic (PLEG): container finished" podID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" exitCode=0 Jan 21 18:34:30 crc kubenswrapper[4808]: I0121 18:34:30.158535 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerDied","Data":"f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0"} Jan 21 18:34:30 crc kubenswrapper[4808]: I0121 18:34:30.158588 4808 scope.go:117] "RemoveContainer" containerID="dd13afd7fd43a6e8f7732443fd57a881906abc4ac14c52babeb4773a1dc24aa8" Jan 21 18:34:30 crc kubenswrapper[4808]: I0121 18:34:30.159384 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:34:30 crc kubenswrapper[4808]: E0121 18:34:30.159695 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:34:41 crc kubenswrapper[4808]: I0121 18:34:41.520663 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:34:41 crc kubenswrapper[4808]: E0121 18:34:41.521779 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:34:52 crc kubenswrapper[4808]: I0121 18:34:52.520980 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:34:52 crc kubenswrapper[4808]: E0121 18:34:52.523383 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:35:03 crc kubenswrapper[4808]: I0121 18:35:03.520697 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:35:03 crc kubenswrapper[4808]: E0121 18:35:03.521524 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:35:18 crc kubenswrapper[4808]: I0121 18:35:18.521295 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:35:18 crc kubenswrapper[4808]: E0121 18:35:18.522183 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:35:31 crc kubenswrapper[4808]: I0121 18:35:31.520505 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:35:31 crc kubenswrapper[4808]: E0121 18:35:31.521136 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:35:44 crc kubenswrapper[4808]: I0121 18:35:44.520326 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:35:44 crc kubenswrapper[4808]: E0121 18:35:44.521000 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:35:55 crc kubenswrapper[4808]: I0121 18:35:55.525689 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:35:55 crc kubenswrapper[4808]: E0121 18:35:55.526341 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:36:10 crc kubenswrapper[4808]: I0121 18:36:10.521211 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:36:10 crc kubenswrapper[4808]: E0121 18:36:10.521705 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:36:22 crc kubenswrapper[4808]: I0121 18:36:22.521020 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:36:22 crc kubenswrapper[4808]: E0121 18:36:22.521649 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:36:37 crc kubenswrapper[4808]: I0121 18:36:37.520729 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:36:37 crc kubenswrapper[4808]: E0121 18:36:37.521505 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:36:48 crc kubenswrapper[4808]: I0121 18:36:48.520476 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:36:48 crc kubenswrapper[4808]: E0121 18:36:48.521359 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:36:59 crc kubenswrapper[4808]: I0121 18:36:59.521946 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:36:59 crc kubenswrapper[4808]: E0121 18:36:59.522716 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:37:12 crc kubenswrapper[4808]: I0121 18:37:12.521265 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:37:12 crc kubenswrapper[4808]: E0121 18:37:12.522015 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:37:25 crc kubenswrapper[4808]: I0121 18:37:25.523841 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:37:25 crc kubenswrapper[4808]: E0121 18:37:25.524550 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:37:36 crc kubenswrapper[4808]: I0121 18:37:36.520751 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:37:36 crc kubenswrapper[4808]: E0121 18:37:36.521631 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:37:50 crc kubenswrapper[4808]: I0121 18:37:50.522389 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:37:50 crc kubenswrapper[4808]: E0121 18:37:50.523016 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:38:04 crc kubenswrapper[4808]: I0121 18:38:04.520638 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:38:04 crc kubenswrapper[4808]: E0121 18:38:04.521626 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:38:07 crc kubenswrapper[4808]: I0121 18:38:07.236430 4808 scope.go:117] "RemoveContainer" containerID="027c60e4f90074860187e100e9915c3d34d362d276264049bf13c84062869f42" Jan 21 18:38:07 crc kubenswrapper[4808]: I0121 18:38:07.263756 4808 scope.go:117] "RemoveContainer" containerID="c3af51d207d215ec7b88f821fd62556b6a7e67b7a0f7e9ae95d9caac4396b3ee" Jan 21 18:38:07 crc kubenswrapper[4808]: I0121 18:38:07.286656 4808 scope.go:117] "RemoveContainer" containerID="f662727a1fe5f11908a3627b1827ee843587083abb4bad7930ea9ac8f257ecfb" Jan 21 18:38:16 crc kubenswrapper[4808]: I0121 18:38:16.520665 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:38:16 crc kubenswrapper[4808]: E0121 18:38:16.521119 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:38:28 crc kubenswrapper[4808]: I0121 18:38:28.521891 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:38:28 crc kubenswrapper[4808]: E0121 18:38:28.522444 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:38:41 crc kubenswrapper[4808]: I0121 18:38:41.524623 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:38:41 crc kubenswrapper[4808]: E0121 18:38:41.525547 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:38:53 crc kubenswrapper[4808]: I0121 18:38:53.525131 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:38:53 crc kubenswrapper[4808]: E0121 18:38:53.538896 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:39:06 crc kubenswrapper[4808]: I0121 18:39:06.520858 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:39:06 crc kubenswrapper[4808]: E0121 18:39:06.521496 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:39:19 crc kubenswrapper[4808]: I0121 18:39:19.523788 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:39:19 crc kubenswrapper[4808]: E0121 18:39:19.524502 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:39:32 crc kubenswrapper[4808]: I0121 18:39:32.523140 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:39:33 crc kubenswrapper[4808]: I0121 18:39:33.196436 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerStarted","Data":"a18080ca9142975beeb9b6bd295d2bb48d20c10aa2b2c3996a753b10e7cad3fe"} Jan 21 18:40:51 crc kubenswrapper[4808]: I0121 18:40:51.551714 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-b6xp9"] Jan 21 18:40:51 crc kubenswrapper[4808]: E0121 18:40:51.552689 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2157473d-7ad3-45a3-8215-dbf88187ae6c" containerName="registry-server" Jan 21 18:40:51 crc kubenswrapper[4808]: I0121 18:40:51.552708 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="2157473d-7ad3-45a3-8215-dbf88187ae6c" containerName="registry-server" Jan 21 18:40:51 crc kubenswrapper[4808]: E0121 18:40:51.552737 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2157473d-7ad3-45a3-8215-dbf88187ae6c" containerName="extract-content" Jan 21 18:40:51 crc kubenswrapper[4808]: I0121 18:40:51.552745 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="2157473d-7ad3-45a3-8215-dbf88187ae6c" containerName="extract-content" Jan 21 18:40:51 crc kubenswrapper[4808]: E0121 18:40:51.552763 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2157473d-7ad3-45a3-8215-dbf88187ae6c" containerName="extract-utilities" Jan 21 18:40:51 crc kubenswrapper[4808]: I0121 18:40:51.552770 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="2157473d-7ad3-45a3-8215-dbf88187ae6c" containerName="extract-utilities" Jan 21 18:40:51 crc kubenswrapper[4808]: I0121 18:40:51.552912 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="2157473d-7ad3-45a3-8215-dbf88187ae6c" containerName="registry-server" Jan 21 18:40:51 crc kubenswrapper[4808]: I0121 18:40:51.553949 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b6xp9" Jan 21 18:40:51 crc kubenswrapper[4808]: I0121 18:40:51.565260 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b6xp9"] Jan 21 18:40:51 crc kubenswrapper[4808]: I0121 18:40:51.745816 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8421be77-134a-4b20-9281-7ae97d604a9d-catalog-content\") pod \"redhat-operators-b6xp9\" (UID: \"8421be77-134a-4b20-9281-7ae97d604a9d\") " pod="openshift-marketplace/redhat-operators-b6xp9" Jan 21 18:40:51 crc kubenswrapper[4808]: I0121 18:40:51.745897 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8421be77-134a-4b20-9281-7ae97d604a9d-utilities\") pod \"redhat-operators-b6xp9\" (UID: \"8421be77-134a-4b20-9281-7ae97d604a9d\") " pod="openshift-marketplace/redhat-operators-b6xp9" Jan 21 18:40:51 crc kubenswrapper[4808]: I0121 18:40:51.745959 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkc2h\" (UniqueName: \"kubernetes.io/projected/8421be77-134a-4b20-9281-7ae97d604a9d-kube-api-access-zkc2h\") pod \"redhat-operators-b6xp9\" (UID: \"8421be77-134a-4b20-9281-7ae97d604a9d\") " pod="openshift-marketplace/redhat-operators-b6xp9" Jan 21 18:40:51 crc kubenswrapper[4808]: I0121 18:40:51.847419 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8421be77-134a-4b20-9281-7ae97d604a9d-catalog-content\") pod \"redhat-operators-b6xp9\" (UID: \"8421be77-134a-4b20-9281-7ae97d604a9d\") " pod="openshift-marketplace/redhat-operators-b6xp9" Jan 21 18:40:51 crc kubenswrapper[4808]: I0121 18:40:51.847510 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8421be77-134a-4b20-9281-7ae97d604a9d-utilities\") pod \"redhat-operators-b6xp9\" (UID: \"8421be77-134a-4b20-9281-7ae97d604a9d\") " pod="openshift-marketplace/redhat-operators-b6xp9" Jan 21 18:40:51 crc kubenswrapper[4808]: I0121 18:40:51.847529 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8421be77-134a-4b20-9281-7ae97d604a9d-catalog-content\") pod \"redhat-operators-b6xp9\" (UID: \"8421be77-134a-4b20-9281-7ae97d604a9d\") " pod="openshift-marketplace/redhat-operators-b6xp9" Jan 21 18:40:51 crc kubenswrapper[4808]: I0121 18:40:51.847657 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkc2h\" (UniqueName: \"kubernetes.io/projected/8421be77-134a-4b20-9281-7ae97d604a9d-kube-api-access-zkc2h\") pod \"redhat-operators-b6xp9\" (UID: \"8421be77-134a-4b20-9281-7ae97d604a9d\") " pod="openshift-marketplace/redhat-operators-b6xp9" Jan 21 18:40:51 crc kubenswrapper[4808]: I0121 18:40:51.847905 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8421be77-134a-4b20-9281-7ae97d604a9d-utilities\") pod \"redhat-operators-b6xp9\" (UID: \"8421be77-134a-4b20-9281-7ae97d604a9d\") " pod="openshift-marketplace/redhat-operators-b6xp9" Jan 21 18:40:51 crc kubenswrapper[4808]: I0121 18:40:51.872454 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkc2h\" (UniqueName: \"kubernetes.io/projected/8421be77-134a-4b20-9281-7ae97d604a9d-kube-api-access-zkc2h\") pod \"redhat-operators-b6xp9\" (UID: \"8421be77-134a-4b20-9281-7ae97d604a9d\") " pod="openshift-marketplace/redhat-operators-b6xp9" Jan 21 18:40:51 crc kubenswrapper[4808]: I0121 18:40:51.878156 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b6xp9" Jan 21 18:40:52 crc kubenswrapper[4808]: I0121 18:40:52.095622 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b6xp9"] Jan 21 18:40:52 crc kubenswrapper[4808]: I0121 18:40:52.725265 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6xp9" event={"ID":"8421be77-134a-4b20-9281-7ae97d604a9d","Type":"ContainerStarted","Data":"069fdcdb2578166033137d9ba76a0e6fd64d3104492cc0761c32128cccf7bd18"} Jan 21 18:40:53 crc kubenswrapper[4808]: I0121 18:40:53.733200 4808 generic.go:334] "Generic (PLEG): container finished" podID="8421be77-134a-4b20-9281-7ae97d604a9d" containerID="4a9e2262a001b03c0c5ad7e7f2d8cfd0ae3a752d761bad344738c457cd38bbd8" exitCode=0 Jan 21 18:40:53 crc kubenswrapper[4808]: I0121 18:40:53.733289 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6xp9" event={"ID":"8421be77-134a-4b20-9281-7ae97d604a9d","Type":"ContainerDied","Data":"4a9e2262a001b03c0c5ad7e7f2d8cfd0ae3a752d761bad344738c457cd38bbd8"} Jan 21 18:40:53 crc kubenswrapper[4808]: I0121 18:40:53.736487 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 18:40:55 crc kubenswrapper[4808]: I0121 18:40:55.745938 4808 generic.go:334] "Generic (PLEG): container finished" podID="8421be77-134a-4b20-9281-7ae97d604a9d" containerID="5e0c3b837c2b53f0fce15c17b40ddb2bf1204a433a748806e1174a1f4f386ba8" exitCode=0 Jan 21 18:40:55 crc kubenswrapper[4808]: I0121 18:40:55.746118 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6xp9" event={"ID":"8421be77-134a-4b20-9281-7ae97d604a9d","Type":"ContainerDied","Data":"5e0c3b837c2b53f0fce15c17b40ddb2bf1204a433a748806e1174a1f4f386ba8"} Jan 21 18:40:56 crc kubenswrapper[4808]: I0121 18:40:56.754121 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6xp9" event={"ID":"8421be77-134a-4b20-9281-7ae97d604a9d","Type":"ContainerStarted","Data":"fd39a93530ada8d773de596f5fe7a3cf4cbaeace2c9fabef27753861445de060"} Jan 21 18:40:56 crc kubenswrapper[4808]: I0121 18:40:56.775009 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-b6xp9" podStartSLOduration=3.371910851 podStartE2EDuration="5.774990386s" podCreationTimestamp="2026-01-21 18:40:51 +0000 UTC" firstStartedPulling="2026-01-21 18:40:53.735873219 +0000 UTC m=+2688.466384114" lastFinishedPulling="2026-01-21 18:40:56.138952764 +0000 UTC m=+2690.869463649" observedRunningTime="2026-01-21 18:40:56.77151891 +0000 UTC m=+2691.502029815" watchObservedRunningTime="2026-01-21 18:40:56.774990386 +0000 UTC m=+2691.505501271" Jan 21 18:41:01 crc kubenswrapper[4808]: I0121 18:41:01.879703 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-b6xp9" Jan 21 18:41:01 crc kubenswrapper[4808]: I0121 18:41:01.880017 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-b6xp9" Jan 21 18:41:02 crc kubenswrapper[4808]: I0121 18:41:02.925790 4808 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-b6xp9" podUID="8421be77-134a-4b20-9281-7ae97d604a9d" containerName="registry-server" probeResult="failure" output=< Jan 21 18:41:02 crc kubenswrapper[4808]: timeout: failed to connect service ":50051" within 1s Jan 21 18:41:02 crc kubenswrapper[4808]: > Jan 21 18:41:11 crc kubenswrapper[4808]: I0121 18:41:11.924501 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-b6xp9" Jan 21 18:41:11 crc kubenswrapper[4808]: I0121 18:41:11.971224 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-b6xp9" Jan 21 18:41:12 crc kubenswrapper[4808]: I0121 18:41:12.162806 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b6xp9"] Jan 21 18:41:13 crc kubenswrapper[4808]: I0121 18:41:13.876199 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-b6xp9" podUID="8421be77-134a-4b20-9281-7ae97d604a9d" containerName="registry-server" containerID="cri-o://fd39a93530ada8d773de596f5fe7a3cf4cbaeace2c9fabef27753861445de060" gracePeriod=2 Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.251205 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b6xp9" Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.386836 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8421be77-134a-4b20-9281-7ae97d604a9d-utilities\") pod \"8421be77-134a-4b20-9281-7ae97d604a9d\" (UID: \"8421be77-134a-4b20-9281-7ae97d604a9d\") " Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.387102 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkc2h\" (UniqueName: \"kubernetes.io/projected/8421be77-134a-4b20-9281-7ae97d604a9d-kube-api-access-zkc2h\") pod \"8421be77-134a-4b20-9281-7ae97d604a9d\" (UID: \"8421be77-134a-4b20-9281-7ae97d604a9d\") " Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.387137 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8421be77-134a-4b20-9281-7ae97d604a9d-catalog-content\") pod \"8421be77-134a-4b20-9281-7ae97d604a9d\" (UID: \"8421be77-134a-4b20-9281-7ae97d604a9d\") " Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.388280 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8421be77-134a-4b20-9281-7ae97d604a9d-utilities" (OuterVolumeSpecName: "utilities") pod "8421be77-134a-4b20-9281-7ae97d604a9d" (UID: "8421be77-134a-4b20-9281-7ae97d604a9d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.392309 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8421be77-134a-4b20-9281-7ae97d604a9d-kube-api-access-zkc2h" (OuterVolumeSpecName: "kube-api-access-zkc2h") pod "8421be77-134a-4b20-9281-7ae97d604a9d" (UID: "8421be77-134a-4b20-9281-7ae97d604a9d"). InnerVolumeSpecName "kube-api-access-zkc2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.488500 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkc2h\" (UniqueName: \"kubernetes.io/projected/8421be77-134a-4b20-9281-7ae97d604a9d-kube-api-access-zkc2h\") on node \"crc\" DevicePath \"\"" Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.488544 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8421be77-134a-4b20-9281-7ae97d604a9d-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.516636 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8421be77-134a-4b20-9281-7ae97d604a9d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8421be77-134a-4b20-9281-7ae97d604a9d" (UID: "8421be77-134a-4b20-9281-7ae97d604a9d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.589334 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8421be77-134a-4b20-9281-7ae97d604a9d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.886063 4808 generic.go:334] "Generic (PLEG): container finished" podID="8421be77-134a-4b20-9281-7ae97d604a9d" containerID="fd39a93530ada8d773de596f5fe7a3cf4cbaeace2c9fabef27753861445de060" exitCode=0 Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.886123 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6xp9" event={"ID":"8421be77-134a-4b20-9281-7ae97d604a9d","Type":"ContainerDied","Data":"fd39a93530ada8d773de596f5fe7a3cf4cbaeace2c9fabef27753861445de060"} Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.886134 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b6xp9" Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.886171 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6xp9" event={"ID":"8421be77-134a-4b20-9281-7ae97d604a9d","Type":"ContainerDied","Data":"069fdcdb2578166033137d9ba76a0e6fd64d3104492cc0761c32128cccf7bd18"} Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.886195 4808 scope.go:117] "RemoveContainer" containerID="fd39a93530ada8d773de596f5fe7a3cf4cbaeace2c9fabef27753861445de060" Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.924903 4808 scope.go:117] "RemoveContainer" containerID="5e0c3b837c2b53f0fce15c17b40ddb2bf1204a433a748806e1174a1f4f386ba8" Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.929237 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b6xp9"] Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.936201 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-b6xp9"] Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.952705 4808 scope.go:117] "RemoveContainer" containerID="4a9e2262a001b03c0c5ad7e7f2d8cfd0ae3a752d761bad344738c457cd38bbd8" Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.969535 4808 scope.go:117] "RemoveContainer" containerID="fd39a93530ada8d773de596f5fe7a3cf4cbaeace2c9fabef27753861445de060" Jan 21 18:41:14 crc kubenswrapper[4808]: E0121 18:41:14.970091 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd39a93530ada8d773de596f5fe7a3cf4cbaeace2c9fabef27753861445de060\": container with ID starting with fd39a93530ada8d773de596f5fe7a3cf4cbaeace2c9fabef27753861445de060 not found: ID does not exist" containerID="fd39a93530ada8d773de596f5fe7a3cf4cbaeace2c9fabef27753861445de060" Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.970148 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd39a93530ada8d773de596f5fe7a3cf4cbaeace2c9fabef27753861445de060"} err="failed to get container status \"fd39a93530ada8d773de596f5fe7a3cf4cbaeace2c9fabef27753861445de060\": rpc error: code = NotFound desc = could not find container \"fd39a93530ada8d773de596f5fe7a3cf4cbaeace2c9fabef27753861445de060\": container with ID starting with fd39a93530ada8d773de596f5fe7a3cf4cbaeace2c9fabef27753861445de060 not found: ID does not exist" Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.970187 4808 scope.go:117] "RemoveContainer" containerID="5e0c3b837c2b53f0fce15c17b40ddb2bf1204a433a748806e1174a1f4f386ba8" Jan 21 18:41:14 crc kubenswrapper[4808]: E0121 18:41:14.970948 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e0c3b837c2b53f0fce15c17b40ddb2bf1204a433a748806e1174a1f4f386ba8\": container with ID starting with 5e0c3b837c2b53f0fce15c17b40ddb2bf1204a433a748806e1174a1f4f386ba8 not found: ID does not exist" containerID="5e0c3b837c2b53f0fce15c17b40ddb2bf1204a433a748806e1174a1f4f386ba8" Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.970987 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e0c3b837c2b53f0fce15c17b40ddb2bf1204a433a748806e1174a1f4f386ba8"} err="failed to get container status \"5e0c3b837c2b53f0fce15c17b40ddb2bf1204a433a748806e1174a1f4f386ba8\": rpc error: code = NotFound desc = could not find container \"5e0c3b837c2b53f0fce15c17b40ddb2bf1204a433a748806e1174a1f4f386ba8\": container with ID starting with 5e0c3b837c2b53f0fce15c17b40ddb2bf1204a433a748806e1174a1f4f386ba8 not found: ID does not exist" Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.971015 4808 scope.go:117] "RemoveContainer" containerID="4a9e2262a001b03c0c5ad7e7f2d8cfd0ae3a752d761bad344738c457cd38bbd8" Jan 21 18:41:14 crc kubenswrapper[4808]: E0121 18:41:14.971408 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a9e2262a001b03c0c5ad7e7f2d8cfd0ae3a752d761bad344738c457cd38bbd8\": container with ID starting with 4a9e2262a001b03c0c5ad7e7f2d8cfd0ae3a752d761bad344738c457cd38bbd8 not found: ID does not exist" containerID="4a9e2262a001b03c0c5ad7e7f2d8cfd0ae3a752d761bad344738c457cd38bbd8" Jan 21 18:41:14 crc kubenswrapper[4808]: I0121 18:41:14.971455 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a9e2262a001b03c0c5ad7e7f2d8cfd0ae3a752d761bad344738c457cd38bbd8"} err="failed to get container status \"4a9e2262a001b03c0c5ad7e7f2d8cfd0ae3a752d761bad344738c457cd38bbd8\": rpc error: code = NotFound desc = could not find container \"4a9e2262a001b03c0c5ad7e7f2d8cfd0ae3a752d761bad344738c457cd38bbd8\": container with ID starting with 4a9e2262a001b03c0c5ad7e7f2d8cfd0ae3a752d761bad344738c457cd38bbd8 not found: ID does not exist" Jan 21 18:41:15 crc kubenswrapper[4808]: I0121 18:41:15.529661 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8421be77-134a-4b20-9281-7ae97d604a9d" path="/var/lib/kubelet/pods/8421be77-134a-4b20-9281-7ae97d604a9d/volumes" Jan 21 18:41:51 crc kubenswrapper[4808]: I0121 18:41:51.502043 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-w4rvt"] Jan 21 18:41:51 crc kubenswrapper[4808]: E0121 18:41:51.502986 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8421be77-134a-4b20-9281-7ae97d604a9d" containerName="extract-utilities" Jan 21 18:41:51 crc kubenswrapper[4808]: I0121 18:41:51.503005 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8421be77-134a-4b20-9281-7ae97d604a9d" containerName="extract-utilities" Jan 21 18:41:51 crc kubenswrapper[4808]: E0121 18:41:51.503038 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8421be77-134a-4b20-9281-7ae97d604a9d" containerName="registry-server" Jan 21 18:41:51 crc kubenswrapper[4808]: I0121 18:41:51.503046 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8421be77-134a-4b20-9281-7ae97d604a9d" containerName="registry-server" Jan 21 18:41:51 crc kubenswrapper[4808]: E0121 18:41:51.503066 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8421be77-134a-4b20-9281-7ae97d604a9d" containerName="extract-content" Jan 21 18:41:51 crc kubenswrapper[4808]: I0121 18:41:51.503075 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="8421be77-134a-4b20-9281-7ae97d604a9d" containerName="extract-content" Jan 21 18:41:51 crc kubenswrapper[4808]: I0121 18:41:51.503225 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="8421be77-134a-4b20-9281-7ae97d604a9d" containerName="registry-server" Jan 21 18:41:51 crc kubenswrapper[4808]: I0121 18:41:51.504312 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w4rvt" Jan 21 18:41:51 crc kubenswrapper[4808]: I0121 18:41:51.512409 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w4rvt"] Jan 21 18:41:51 crc kubenswrapper[4808]: I0121 18:41:51.555401 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db9d296d-26c8-4504-a8f9-f40a7c46fafc-utilities\") pod \"certified-operators-w4rvt\" (UID: \"db9d296d-26c8-4504-a8f9-f40a7c46fafc\") " pod="openshift-marketplace/certified-operators-w4rvt" Jan 21 18:41:51 crc kubenswrapper[4808]: I0121 18:41:51.555574 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db9d296d-26c8-4504-a8f9-f40a7c46fafc-catalog-content\") pod \"certified-operators-w4rvt\" (UID: \"db9d296d-26c8-4504-a8f9-f40a7c46fafc\") " pod="openshift-marketplace/certified-operators-w4rvt" Jan 21 18:41:51 crc kubenswrapper[4808]: I0121 18:41:51.555791 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc4bt\" (UniqueName: \"kubernetes.io/projected/db9d296d-26c8-4504-a8f9-f40a7c46fafc-kube-api-access-fc4bt\") pod \"certified-operators-w4rvt\" (UID: \"db9d296d-26c8-4504-a8f9-f40a7c46fafc\") " pod="openshift-marketplace/certified-operators-w4rvt" Jan 21 18:41:51 crc kubenswrapper[4808]: I0121 18:41:51.657447 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc4bt\" (UniqueName: \"kubernetes.io/projected/db9d296d-26c8-4504-a8f9-f40a7c46fafc-kube-api-access-fc4bt\") pod \"certified-operators-w4rvt\" (UID: \"db9d296d-26c8-4504-a8f9-f40a7c46fafc\") " pod="openshift-marketplace/certified-operators-w4rvt" Jan 21 18:41:51 crc kubenswrapper[4808]: I0121 18:41:51.657567 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db9d296d-26c8-4504-a8f9-f40a7c46fafc-utilities\") pod \"certified-operators-w4rvt\" (UID: \"db9d296d-26c8-4504-a8f9-f40a7c46fafc\") " pod="openshift-marketplace/certified-operators-w4rvt" Jan 21 18:41:51 crc kubenswrapper[4808]: I0121 18:41:51.657595 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db9d296d-26c8-4504-a8f9-f40a7c46fafc-catalog-content\") pod \"certified-operators-w4rvt\" (UID: \"db9d296d-26c8-4504-a8f9-f40a7c46fafc\") " pod="openshift-marketplace/certified-operators-w4rvt" Jan 21 18:41:51 crc kubenswrapper[4808]: I0121 18:41:51.658010 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db9d296d-26c8-4504-a8f9-f40a7c46fafc-utilities\") pod \"certified-operators-w4rvt\" (UID: \"db9d296d-26c8-4504-a8f9-f40a7c46fafc\") " pod="openshift-marketplace/certified-operators-w4rvt" Jan 21 18:41:51 crc kubenswrapper[4808]: I0121 18:41:51.658160 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db9d296d-26c8-4504-a8f9-f40a7c46fafc-catalog-content\") pod \"certified-operators-w4rvt\" (UID: \"db9d296d-26c8-4504-a8f9-f40a7c46fafc\") " pod="openshift-marketplace/certified-operators-w4rvt" Jan 21 18:41:51 crc kubenswrapper[4808]: I0121 18:41:51.684992 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc4bt\" (UniqueName: \"kubernetes.io/projected/db9d296d-26c8-4504-a8f9-f40a7c46fafc-kube-api-access-fc4bt\") pod \"certified-operators-w4rvt\" (UID: \"db9d296d-26c8-4504-a8f9-f40a7c46fafc\") " pod="openshift-marketplace/certified-operators-w4rvt" Jan 21 18:41:51 crc kubenswrapper[4808]: I0121 18:41:51.823455 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w4rvt" Jan 21 18:41:52 crc kubenswrapper[4808]: I0121 18:41:52.275025 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w4rvt"] Jan 21 18:41:53 crc kubenswrapper[4808]: I0121 18:41:53.115214 4808 generic.go:334] "Generic (PLEG): container finished" podID="db9d296d-26c8-4504-a8f9-f40a7c46fafc" containerID="556006f28d507640f2f2aa54cea8297a0cd035ff5ae3867443ecccc03300e5b8" exitCode=0 Jan 21 18:41:53 crc kubenswrapper[4808]: I0121 18:41:53.115295 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w4rvt" event={"ID":"db9d296d-26c8-4504-a8f9-f40a7c46fafc","Type":"ContainerDied","Data":"556006f28d507640f2f2aa54cea8297a0cd035ff5ae3867443ecccc03300e5b8"} Jan 21 18:41:53 crc kubenswrapper[4808]: I0121 18:41:53.115335 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w4rvt" event={"ID":"db9d296d-26c8-4504-a8f9-f40a7c46fafc","Type":"ContainerStarted","Data":"f1624abcee43de31e451790468ab89b86d13b44831f0cf3c5692549f48e025de"} Jan 21 18:41:54 crc kubenswrapper[4808]: I0121 18:41:54.123509 4808 generic.go:334] "Generic (PLEG): container finished" podID="db9d296d-26c8-4504-a8f9-f40a7c46fafc" containerID="11faf7ee9e9afcfdfb782abaf81e9972e1f2976c0a5edf5bc19b1ee80469fdf4" exitCode=0 Jan 21 18:41:54 crc kubenswrapper[4808]: I0121 18:41:54.123719 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w4rvt" event={"ID":"db9d296d-26c8-4504-a8f9-f40a7c46fafc","Type":"ContainerDied","Data":"11faf7ee9e9afcfdfb782abaf81e9972e1f2976c0a5edf5bc19b1ee80469fdf4"} Jan 21 18:41:55 crc kubenswrapper[4808]: I0121 18:41:55.131793 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w4rvt" event={"ID":"db9d296d-26c8-4504-a8f9-f40a7c46fafc","Type":"ContainerStarted","Data":"d7023515bab47ebcd2989574eb51bc5861307c9629f7256adef0ba38e9d4678c"} Jan 21 18:41:55 crc kubenswrapper[4808]: I0121 18:41:55.153684 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-w4rvt" podStartSLOduration=2.732254356 podStartE2EDuration="4.153666901s" podCreationTimestamp="2026-01-21 18:41:51 +0000 UTC" firstStartedPulling="2026-01-21 18:41:53.117031206 +0000 UTC m=+2747.847542101" lastFinishedPulling="2026-01-21 18:41:54.538443761 +0000 UTC m=+2749.268954646" observedRunningTime="2026-01-21 18:41:55.148092794 +0000 UTC m=+2749.878603699" watchObservedRunningTime="2026-01-21 18:41:55.153666901 +0000 UTC m=+2749.884177786" Jan 21 18:41:59 crc kubenswrapper[4808]: I0121 18:41:59.599310 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:41:59 crc kubenswrapper[4808]: I0121 18:41:59.599862 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:42:01 crc kubenswrapper[4808]: I0121 18:42:01.823880 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-w4rvt" Jan 21 18:42:01 crc kubenswrapper[4808]: I0121 18:42:01.823929 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-w4rvt" Jan 21 18:42:01 crc kubenswrapper[4808]: I0121 18:42:01.866881 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-w4rvt" Jan 21 18:42:02 crc kubenswrapper[4808]: I0121 18:42:02.218668 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-w4rvt" Jan 21 18:42:02 crc kubenswrapper[4808]: I0121 18:42:02.267115 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w4rvt"] Jan 21 18:42:04 crc kubenswrapper[4808]: I0121 18:42:04.197041 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-w4rvt" podUID="db9d296d-26c8-4504-a8f9-f40a7c46fafc" containerName="registry-server" containerID="cri-o://d7023515bab47ebcd2989574eb51bc5861307c9629f7256adef0ba38e9d4678c" gracePeriod=2 Jan 21 18:42:05 crc kubenswrapper[4808]: I0121 18:42:05.205069 4808 generic.go:334] "Generic (PLEG): container finished" podID="db9d296d-26c8-4504-a8f9-f40a7c46fafc" containerID="d7023515bab47ebcd2989574eb51bc5861307c9629f7256adef0ba38e9d4678c" exitCode=0 Jan 21 18:42:05 crc kubenswrapper[4808]: I0121 18:42:05.205382 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w4rvt" event={"ID":"db9d296d-26c8-4504-a8f9-f40a7c46fafc","Type":"ContainerDied","Data":"d7023515bab47ebcd2989574eb51bc5861307c9629f7256adef0ba38e9d4678c"} Jan 21 18:42:05 crc kubenswrapper[4808]: I0121 18:42:05.669534 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w4rvt" Jan 21 18:42:05 crc kubenswrapper[4808]: I0121 18:42:05.746956 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db9d296d-26c8-4504-a8f9-f40a7c46fafc-catalog-content\") pod \"db9d296d-26c8-4504-a8f9-f40a7c46fafc\" (UID: \"db9d296d-26c8-4504-a8f9-f40a7c46fafc\") " Jan 21 18:42:05 crc kubenswrapper[4808]: I0121 18:42:05.747285 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc4bt\" (UniqueName: \"kubernetes.io/projected/db9d296d-26c8-4504-a8f9-f40a7c46fafc-kube-api-access-fc4bt\") pod \"db9d296d-26c8-4504-a8f9-f40a7c46fafc\" (UID: \"db9d296d-26c8-4504-a8f9-f40a7c46fafc\") " Jan 21 18:42:05 crc kubenswrapper[4808]: I0121 18:42:05.747415 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db9d296d-26c8-4504-a8f9-f40a7c46fafc-utilities\") pod \"db9d296d-26c8-4504-a8f9-f40a7c46fafc\" (UID: \"db9d296d-26c8-4504-a8f9-f40a7c46fafc\") " Jan 21 18:42:05 crc kubenswrapper[4808]: I0121 18:42:05.748856 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db9d296d-26c8-4504-a8f9-f40a7c46fafc-utilities" (OuterVolumeSpecName: "utilities") pod "db9d296d-26c8-4504-a8f9-f40a7c46fafc" (UID: "db9d296d-26c8-4504-a8f9-f40a7c46fafc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:42:05 crc kubenswrapper[4808]: I0121 18:42:05.754638 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db9d296d-26c8-4504-a8f9-f40a7c46fafc-kube-api-access-fc4bt" (OuterVolumeSpecName: "kube-api-access-fc4bt") pod "db9d296d-26c8-4504-a8f9-f40a7c46fafc" (UID: "db9d296d-26c8-4504-a8f9-f40a7c46fafc"). InnerVolumeSpecName "kube-api-access-fc4bt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:42:05 crc kubenswrapper[4808]: I0121 18:42:05.811676 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db9d296d-26c8-4504-a8f9-f40a7c46fafc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db9d296d-26c8-4504-a8f9-f40a7c46fafc" (UID: "db9d296d-26c8-4504-a8f9-f40a7c46fafc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:42:05 crc kubenswrapper[4808]: I0121 18:42:05.849094 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db9d296d-26c8-4504-a8f9-f40a7c46fafc-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:42:05 crc kubenswrapper[4808]: I0121 18:42:05.849142 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db9d296d-26c8-4504-a8f9-f40a7c46fafc-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:42:05 crc kubenswrapper[4808]: I0121 18:42:05.849161 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc4bt\" (UniqueName: \"kubernetes.io/projected/db9d296d-26c8-4504-a8f9-f40a7c46fafc-kube-api-access-fc4bt\") on node \"crc\" DevicePath \"\"" Jan 21 18:42:06 crc kubenswrapper[4808]: I0121 18:42:06.213891 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w4rvt" event={"ID":"db9d296d-26c8-4504-a8f9-f40a7c46fafc","Type":"ContainerDied","Data":"f1624abcee43de31e451790468ab89b86d13b44831f0cf3c5692549f48e025de"} Jan 21 18:42:06 crc kubenswrapper[4808]: I0121 18:42:06.214013 4808 scope.go:117] "RemoveContainer" containerID="d7023515bab47ebcd2989574eb51bc5861307c9629f7256adef0ba38e9d4678c" Jan 21 18:42:06 crc kubenswrapper[4808]: I0121 18:42:06.214135 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w4rvt" Jan 21 18:42:06 crc kubenswrapper[4808]: I0121 18:42:06.250222 4808 scope.go:117] "RemoveContainer" containerID="11faf7ee9e9afcfdfb782abaf81e9972e1f2976c0a5edf5bc19b1ee80469fdf4" Jan 21 18:42:06 crc kubenswrapper[4808]: I0121 18:42:06.256830 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w4rvt"] Jan 21 18:42:06 crc kubenswrapper[4808]: I0121 18:42:06.262599 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-w4rvt"] Jan 21 18:42:06 crc kubenswrapper[4808]: I0121 18:42:06.269965 4808 scope.go:117] "RemoveContainer" containerID="556006f28d507640f2f2aa54cea8297a0cd035ff5ae3867443ecccc03300e5b8" Jan 21 18:42:07 crc kubenswrapper[4808]: I0121 18:42:07.530926 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db9d296d-26c8-4504-a8f9-f40a7c46fafc" path="/var/lib/kubelet/pods/db9d296d-26c8-4504-a8f9-f40a7c46fafc/volumes" Jan 21 18:42:20 crc kubenswrapper[4808]: I0121 18:42:20.429366 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ngqwn"] Jan 21 18:42:20 crc kubenswrapper[4808]: E0121 18:42:20.430192 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db9d296d-26c8-4504-a8f9-f40a7c46fafc" containerName="extract-utilities" Jan 21 18:42:20 crc kubenswrapper[4808]: I0121 18:42:20.430208 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="db9d296d-26c8-4504-a8f9-f40a7c46fafc" containerName="extract-utilities" Jan 21 18:42:20 crc kubenswrapper[4808]: E0121 18:42:20.430223 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db9d296d-26c8-4504-a8f9-f40a7c46fafc" containerName="extract-content" Jan 21 18:42:20 crc kubenswrapper[4808]: I0121 18:42:20.430231 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="db9d296d-26c8-4504-a8f9-f40a7c46fafc" containerName="extract-content" Jan 21 18:42:20 crc kubenswrapper[4808]: E0121 18:42:20.430265 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db9d296d-26c8-4504-a8f9-f40a7c46fafc" containerName="registry-server" Jan 21 18:42:20 crc kubenswrapper[4808]: I0121 18:42:20.430272 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="db9d296d-26c8-4504-a8f9-f40a7c46fafc" containerName="registry-server" Jan 21 18:42:20 crc kubenswrapper[4808]: I0121 18:42:20.430399 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="db9d296d-26c8-4504-a8f9-f40a7c46fafc" containerName="registry-server" Jan 21 18:42:20 crc kubenswrapper[4808]: I0121 18:42:20.431258 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ngqwn" Jan 21 18:42:20 crc kubenswrapper[4808]: I0121 18:42:20.437450 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ngqwn"] Jan 21 18:42:20 crc kubenswrapper[4808]: I0121 18:42:20.467851 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85268\" (UniqueName: \"kubernetes.io/projected/725ad597-130c-4ed2-a3e2-59b39cab1401-kube-api-access-85268\") pod \"community-operators-ngqwn\" (UID: \"725ad597-130c-4ed2-a3e2-59b39cab1401\") " pod="openshift-marketplace/community-operators-ngqwn" Jan 21 18:42:20 crc kubenswrapper[4808]: I0121 18:42:20.467918 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/725ad597-130c-4ed2-a3e2-59b39cab1401-utilities\") pod \"community-operators-ngqwn\" (UID: \"725ad597-130c-4ed2-a3e2-59b39cab1401\") " pod="openshift-marketplace/community-operators-ngqwn" Jan 21 18:42:20 crc kubenswrapper[4808]: I0121 18:42:20.468266 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/725ad597-130c-4ed2-a3e2-59b39cab1401-catalog-content\") pod \"community-operators-ngqwn\" (UID: \"725ad597-130c-4ed2-a3e2-59b39cab1401\") " pod="openshift-marketplace/community-operators-ngqwn" Jan 21 18:42:20 crc kubenswrapper[4808]: I0121 18:42:20.569563 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/725ad597-130c-4ed2-a3e2-59b39cab1401-catalog-content\") pod \"community-operators-ngqwn\" (UID: \"725ad597-130c-4ed2-a3e2-59b39cab1401\") " pod="openshift-marketplace/community-operators-ngqwn" Jan 21 18:42:20 crc kubenswrapper[4808]: I0121 18:42:20.569647 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85268\" (UniqueName: \"kubernetes.io/projected/725ad597-130c-4ed2-a3e2-59b39cab1401-kube-api-access-85268\") pod \"community-operators-ngqwn\" (UID: \"725ad597-130c-4ed2-a3e2-59b39cab1401\") " pod="openshift-marketplace/community-operators-ngqwn" Jan 21 18:42:20 crc kubenswrapper[4808]: I0121 18:42:20.569672 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/725ad597-130c-4ed2-a3e2-59b39cab1401-utilities\") pod \"community-operators-ngqwn\" (UID: \"725ad597-130c-4ed2-a3e2-59b39cab1401\") " pod="openshift-marketplace/community-operators-ngqwn" Jan 21 18:42:20 crc kubenswrapper[4808]: I0121 18:42:20.570155 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/725ad597-130c-4ed2-a3e2-59b39cab1401-utilities\") pod \"community-operators-ngqwn\" (UID: \"725ad597-130c-4ed2-a3e2-59b39cab1401\") " pod="openshift-marketplace/community-operators-ngqwn" Jan 21 18:42:20 crc kubenswrapper[4808]: I0121 18:42:20.570223 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/725ad597-130c-4ed2-a3e2-59b39cab1401-catalog-content\") pod \"community-operators-ngqwn\" (UID: \"725ad597-130c-4ed2-a3e2-59b39cab1401\") " pod="openshift-marketplace/community-operators-ngqwn" Jan 21 18:42:20 crc kubenswrapper[4808]: I0121 18:42:20.588819 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85268\" (UniqueName: \"kubernetes.io/projected/725ad597-130c-4ed2-a3e2-59b39cab1401-kube-api-access-85268\") pod \"community-operators-ngqwn\" (UID: \"725ad597-130c-4ed2-a3e2-59b39cab1401\") " pod="openshift-marketplace/community-operators-ngqwn" Jan 21 18:42:20 crc kubenswrapper[4808]: I0121 18:42:20.747643 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ngqwn" Jan 21 18:42:21 crc kubenswrapper[4808]: I0121 18:42:21.238996 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ngqwn"] Jan 21 18:42:21 crc kubenswrapper[4808]: I0121 18:42:21.316516 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ngqwn" event={"ID":"725ad597-130c-4ed2-a3e2-59b39cab1401","Type":"ContainerStarted","Data":"48327ef4092252d3a8b7eaf0065d263c4f56b6ce290972c6ba4f3d75a3ccc89f"} Jan 21 18:42:22 crc kubenswrapper[4808]: I0121 18:42:22.325752 4808 generic.go:334] "Generic (PLEG): container finished" podID="725ad597-130c-4ed2-a3e2-59b39cab1401" containerID="d86fd32d4933f9dc71d92e8e7831850a66873010fc9ca1f9a340c25464d21de9" exitCode=0 Jan 21 18:42:22 crc kubenswrapper[4808]: I0121 18:42:22.325890 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ngqwn" event={"ID":"725ad597-130c-4ed2-a3e2-59b39cab1401","Type":"ContainerDied","Data":"d86fd32d4933f9dc71d92e8e7831850a66873010fc9ca1f9a340c25464d21de9"} Jan 21 18:42:23 crc kubenswrapper[4808]: I0121 18:42:23.336745 4808 generic.go:334] "Generic (PLEG): container finished" podID="725ad597-130c-4ed2-a3e2-59b39cab1401" containerID="7f9ed03361323968181e568801ab4c045e741b7e9a5347b54003a34946a78fb7" exitCode=0 Jan 21 18:42:23 crc kubenswrapper[4808]: I0121 18:42:23.336961 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ngqwn" event={"ID":"725ad597-130c-4ed2-a3e2-59b39cab1401","Type":"ContainerDied","Data":"7f9ed03361323968181e568801ab4c045e741b7e9a5347b54003a34946a78fb7"} Jan 21 18:42:24 crc kubenswrapper[4808]: I0121 18:42:24.346052 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ngqwn" event={"ID":"725ad597-130c-4ed2-a3e2-59b39cab1401","Type":"ContainerStarted","Data":"48c2500be770619fda0f6d3bd6ca0175f1cc6a0d6e30bdae5592f5fbe87f8683"} Jan 21 18:42:24 crc kubenswrapper[4808]: I0121 18:42:24.372216 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ngqwn" podStartSLOduration=2.964657281 podStartE2EDuration="4.372198865s" podCreationTimestamp="2026-01-21 18:42:20 +0000 UTC" firstStartedPulling="2026-01-21 18:42:22.328145789 +0000 UTC m=+2777.058656684" lastFinishedPulling="2026-01-21 18:42:23.735687383 +0000 UTC m=+2778.466198268" observedRunningTime="2026-01-21 18:42:24.36913628 +0000 UTC m=+2779.099647185" watchObservedRunningTime="2026-01-21 18:42:24.372198865 +0000 UTC m=+2779.102709750" Jan 21 18:42:29 crc kubenswrapper[4808]: I0121 18:42:29.599404 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:42:29 crc kubenswrapper[4808]: I0121 18:42:29.600065 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:42:30 crc kubenswrapper[4808]: I0121 18:42:30.748065 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ngqwn" Jan 21 18:42:30 crc kubenswrapper[4808]: I0121 18:42:30.748125 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ngqwn" Jan 21 18:42:30 crc kubenswrapper[4808]: I0121 18:42:30.786194 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ngqwn" Jan 21 18:42:31 crc kubenswrapper[4808]: I0121 18:42:31.432595 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ngqwn" Jan 21 18:42:31 crc kubenswrapper[4808]: I0121 18:42:31.477496 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ngqwn"] Jan 21 18:42:33 crc kubenswrapper[4808]: I0121 18:42:33.402299 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ngqwn" podUID="725ad597-130c-4ed2-a3e2-59b39cab1401" containerName="registry-server" containerID="cri-o://48c2500be770619fda0f6d3bd6ca0175f1cc6a0d6e30bdae5592f5fbe87f8683" gracePeriod=2 Jan 21 18:42:34 crc kubenswrapper[4808]: I0121 18:42:34.411319 4808 generic.go:334] "Generic (PLEG): container finished" podID="725ad597-130c-4ed2-a3e2-59b39cab1401" containerID="48c2500be770619fda0f6d3bd6ca0175f1cc6a0d6e30bdae5592f5fbe87f8683" exitCode=0 Jan 21 18:42:34 crc kubenswrapper[4808]: I0121 18:42:34.411396 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ngqwn" event={"ID":"725ad597-130c-4ed2-a3e2-59b39cab1401","Type":"ContainerDied","Data":"48c2500be770619fda0f6d3bd6ca0175f1cc6a0d6e30bdae5592f5fbe87f8683"} Jan 21 18:42:34 crc kubenswrapper[4808]: I0121 18:42:34.763303 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ngqwn" Jan 21 18:42:34 crc kubenswrapper[4808]: I0121 18:42:34.778972 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/725ad597-130c-4ed2-a3e2-59b39cab1401-catalog-content\") pod \"725ad597-130c-4ed2-a3e2-59b39cab1401\" (UID: \"725ad597-130c-4ed2-a3e2-59b39cab1401\") " Jan 21 18:42:34 crc kubenswrapper[4808]: I0121 18:42:34.779024 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/725ad597-130c-4ed2-a3e2-59b39cab1401-utilities\") pod \"725ad597-130c-4ed2-a3e2-59b39cab1401\" (UID: \"725ad597-130c-4ed2-a3e2-59b39cab1401\") " Jan 21 18:42:34 crc kubenswrapper[4808]: I0121 18:42:34.779994 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/725ad597-130c-4ed2-a3e2-59b39cab1401-utilities" (OuterVolumeSpecName: "utilities") pod "725ad597-130c-4ed2-a3e2-59b39cab1401" (UID: "725ad597-130c-4ed2-a3e2-59b39cab1401"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:42:34 crc kubenswrapper[4808]: I0121 18:42:34.835211 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/725ad597-130c-4ed2-a3e2-59b39cab1401-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "725ad597-130c-4ed2-a3e2-59b39cab1401" (UID: "725ad597-130c-4ed2-a3e2-59b39cab1401"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:42:34 crc kubenswrapper[4808]: I0121 18:42:34.879832 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85268\" (UniqueName: \"kubernetes.io/projected/725ad597-130c-4ed2-a3e2-59b39cab1401-kube-api-access-85268\") pod \"725ad597-130c-4ed2-a3e2-59b39cab1401\" (UID: \"725ad597-130c-4ed2-a3e2-59b39cab1401\") " Jan 21 18:42:34 crc kubenswrapper[4808]: I0121 18:42:34.880146 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/725ad597-130c-4ed2-a3e2-59b39cab1401-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:42:34 crc kubenswrapper[4808]: I0121 18:42:34.880164 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/725ad597-130c-4ed2-a3e2-59b39cab1401-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:42:34 crc kubenswrapper[4808]: I0121 18:42:34.886151 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/725ad597-130c-4ed2-a3e2-59b39cab1401-kube-api-access-85268" (OuterVolumeSpecName: "kube-api-access-85268") pod "725ad597-130c-4ed2-a3e2-59b39cab1401" (UID: "725ad597-130c-4ed2-a3e2-59b39cab1401"). InnerVolumeSpecName "kube-api-access-85268". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:42:34 crc kubenswrapper[4808]: I0121 18:42:34.981613 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85268\" (UniqueName: \"kubernetes.io/projected/725ad597-130c-4ed2-a3e2-59b39cab1401-kube-api-access-85268\") on node \"crc\" DevicePath \"\"" Jan 21 18:42:35 crc kubenswrapper[4808]: I0121 18:42:35.419429 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ngqwn" event={"ID":"725ad597-130c-4ed2-a3e2-59b39cab1401","Type":"ContainerDied","Data":"48327ef4092252d3a8b7eaf0065d263c4f56b6ce290972c6ba4f3d75a3ccc89f"} Jan 21 18:42:35 crc kubenswrapper[4808]: I0121 18:42:35.419484 4808 scope.go:117] "RemoveContainer" containerID="48c2500be770619fda0f6d3bd6ca0175f1cc6a0d6e30bdae5592f5fbe87f8683" Jan 21 18:42:35 crc kubenswrapper[4808]: I0121 18:42:35.419505 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ngqwn" Jan 21 18:42:35 crc kubenswrapper[4808]: I0121 18:42:35.447311 4808 scope.go:117] "RemoveContainer" containerID="7f9ed03361323968181e568801ab4c045e741b7e9a5347b54003a34946a78fb7" Jan 21 18:42:35 crc kubenswrapper[4808]: I0121 18:42:35.455579 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ngqwn"] Jan 21 18:42:35 crc kubenswrapper[4808]: I0121 18:42:35.460795 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ngqwn"] Jan 21 18:42:35 crc kubenswrapper[4808]: I0121 18:42:35.482063 4808 scope.go:117] "RemoveContainer" containerID="d86fd32d4933f9dc71d92e8e7831850a66873010fc9ca1f9a340c25464d21de9" Jan 21 18:42:35 crc kubenswrapper[4808]: I0121 18:42:35.535750 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="725ad597-130c-4ed2-a3e2-59b39cab1401" path="/var/lib/kubelet/pods/725ad597-130c-4ed2-a3e2-59b39cab1401/volumes" Jan 21 18:42:59 crc kubenswrapper[4808]: I0121 18:42:59.599757 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:42:59 crc kubenswrapper[4808]: I0121 18:42:59.601611 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:42:59 crc kubenswrapper[4808]: I0121 18:42:59.601781 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 18:42:59 crc kubenswrapper[4808]: I0121 18:42:59.602461 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a18080ca9142975beeb9b6bd295d2bb48d20c10aa2b2c3996a753b10e7cad3fe"} pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 18:42:59 crc kubenswrapper[4808]: I0121 18:42:59.602601 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" containerID="cri-o://a18080ca9142975beeb9b6bd295d2bb48d20c10aa2b2c3996a753b10e7cad3fe" gracePeriod=600 Jan 21 18:43:00 crc kubenswrapper[4808]: I0121 18:43:00.590453 4808 generic.go:334] "Generic (PLEG): container finished" podID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerID="a18080ca9142975beeb9b6bd295d2bb48d20c10aa2b2c3996a753b10e7cad3fe" exitCode=0 Jan 21 18:43:00 crc kubenswrapper[4808]: I0121 18:43:00.590494 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerDied","Data":"a18080ca9142975beeb9b6bd295d2bb48d20c10aa2b2c3996a753b10e7cad3fe"} Jan 21 18:43:00 crc kubenswrapper[4808]: I0121 18:43:00.590913 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerStarted","Data":"dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72"} Jan 21 18:43:00 crc kubenswrapper[4808]: I0121 18:43:00.590940 4808 scope.go:117] "RemoveContainer" containerID="f9909ed45ed527ae627386cfbbca7ba3e2750edead9cc82717368cdf329942f0" Jan 21 18:44:59 crc kubenswrapper[4808]: I0121 18:44:59.599558 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:44:59 crc kubenswrapper[4808]: I0121 18:44:59.600035 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.142795 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483685-l9l7k"] Jan 21 18:45:00 crc kubenswrapper[4808]: E0121 18:45:00.143133 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="725ad597-130c-4ed2-a3e2-59b39cab1401" containerName="extract-utilities" Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.143151 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="725ad597-130c-4ed2-a3e2-59b39cab1401" containerName="extract-utilities" Jan 21 18:45:00 crc kubenswrapper[4808]: E0121 18:45:00.143164 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="725ad597-130c-4ed2-a3e2-59b39cab1401" containerName="registry-server" Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.143170 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="725ad597-130c-4ed2-a3e2-59b39cab1401" containerName="registry-server" Jan 21 18:45:00 crc kubenswrapper[4808]: E0121 18:45:00.143191 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="725ad597-130c-4ed2-a3e2-59b39cab1401" containerName="extract-content" Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.143198 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="725ad597-130c-4ed2-a3e2-59b39cab1401" containerName="extract-content" Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.143343 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="725ad597-130c-4ed2-a3e2-59b39cab1401" containerName="registry-server" Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.143852 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483685-l9l7k" Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.146078 4808 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.147583 4808 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.159622 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483685-l9l7k"] Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.298436 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/788fb951-0cbb-4138-b9e8-8a6a1bd07128-config-volume\") pod \"collect-profiles-29483685-l9l7k\" (UID: \"788fb951-0cbb-4138-b9e8-8a6a1bd07128\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483685-l9l7k" Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.298504 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/788fb951-0cbb-4138-b9e8-8a6a1bd07128-secret-volume\") pod \"collect-profiles-29483685-l9l7k\" (UID: \"788fb951-0cbb-4138-b9e8-8a6a1bd07128\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483685-l9l7k" Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.298598 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl7m6\" (UniqueName: \"kubernetes.io/projected/788fb951-0cbb-4138-b9e8-8a6a1bd07128-kube-api-access-pl7m6\") pod \"collect-profiles-29483685-l9l7k\" (UID: \"788fb951-0cbb-4138-b9e8-8a6a1bd07128\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483685-l9l7k" Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.400028 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/788fb951-0cbb-4138-b9e8-8a6a1bd07128-config-volume\") pod \"collect-profiles-29483685-l9l7k\" (UID: \"788fb951-0cbb-4138-b9e8-8a6a1bd07128\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483685-l9l7k" Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.400134 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/788fb951-0cbb-4138-b9e8-8a6a1bd07128-secret-volume\") pod \"collect-profiles-29483685-l9l7k\" (UID: \"788fb951-0cbb-4138-b9e8-8a6a1bd07128\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483685-l9l7k" Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.400162 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl7m6\" (UniqueName: \"kubernetes.io/projected/788fb951-0cbb-4138-b9e8-8a6a1bd07128-kube-api-access-pl7m6\") pod \"collect-profiles-29483685-l9l7k\" (UID: \"788fb951-0cbb-4138-b9e8-8a6a1bd07128\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483685-l9l7k" Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.401124 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/788fb951-0cbb-4138-b9e8-8a6a1bd07128-config-volume\") pod \"collect-profiles-29483685-l9l7k\" (UID: \"788fb951-0cbb-4138-b9e8-8a6a1bd07128\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483685-l9l7k" Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.406095 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/788fb951-0cbb-4138-b9e8-8a6a1bd07128-secret-volume\") pod \"collect-profiles-29483685-l9l7k\" (UID: \"788fb951-0cbb-4138-b9e8-8a6a1bd07128\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483685-l9l7k" Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.415814 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl7m6\" (UniqueName: \"kubernetes.io/projected/788fb951-0cbb-4138-b9e8-8a6a1bd07128-kube-api-access-pl7m6\") pod \"collect-profiles-29483685-l9l7k\" (UID: \"788fb951-0cbb-4138-b9e8-8a6a1bd07128\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483685-l9l7k" Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.474100 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483685-l9l7k" Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.684677 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483685-l9l7k"] Jan 21 18:45:00 crc kubenswrapper[4808]: I0121 18:45:00.710101 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483685-l9l7k" event={"ID":"788fb951-0cbb-4138-b9e8-8a6a1bd07128","Type":"ContainerStarted","Data":"719e2442bfc2ea80a88742e652e67d9d6c8ec73bf2a13ced22f6a45f21426512"} Jan 21 18:45:01 crc kubenswrapper[4808]: I0121 18:45:01.720707 4808 generic.go:334] "Generic (PLEG): container finished" podID="788fb951-0cbb-4138-b9e8-8a6a1bd07128" containerID="6fef50a9a1f35f45aa76db07c2c42a71df360125f675aba8a345f2529e145cd2" exitCode=0 Jan 21 18:45:01 crc kubenswrapper[4808]: I0121 18:45:01.720826 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483685-l9l7k" event={"ID":"788fb951-0cbb-4138-b9e8-8a6a1bd07128","Type":"ContainerDied","Data":"6fef50a9a1f35f45aa76db07c2c42a71df360125f675aba8a345f2529e145cd2"} Jan 21 18:45:02 crc kubenswrapper[4808]: I0121 18:45:02.979547 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483685-l9l7k" Jan 21 18:45:03 crc kubenswrapper[4808]: I0121 18:45:03.031798 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pl7m6\" (UniqueName: \"kubernetes.io/projected/788fb951-0cbb-4138-b9e8-8a6a1bd07128-kube-api-access-pl7m6\") pod \"788fb951-0cbb-4138-b9e8-8a6a1bd07128\" (UID: \"788fb951-0cbb-4138-b9e8-8a6a1bd07128\") " Jan 21 18:45:03 crc kubenswrapper[4808]: I0121 18:45:03.032054 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/788fb951-0cbb-4138-b9e8-8a6a1bd07128-config-volume\") pod \"788fb951-0cbb-4138-b9e8-8a6a1bd07128\" (UID: \"788fb951-0cbb-4138-b9e8-8a6a1bd07128\") " Jan 21 18:45:03 crc kubenswrapper[4808]: I0121 18:45:03.032109 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/788fb951-0cbb-4138-b9e8-8a6a1bd07128-secret-volume\") pod \"788fb951-0cbb-4138-b9e8-8a6a1bd07128\" (UID: \"788fb951-0cbb-4138-b9e8-8a6a1bd07128\") " Jan 21 18:45:03 crc kubenswrapper[4808]: I0121 18:45:03.033717 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/788fb951-0cbb-4138-b9e8-8a6a1bd07128-config-volume" (OuterVolumeSpecName: "config-volume") pod "788fb951-0cbb-4138-b9e8-8a6a1bd07128" (UID: "788fb951-0cbb-4138-b9e8-8a6a1bd07128"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 18:45:03 crc kubenswrapper[4808]: I0121 18:45:03.039649 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/788fb951-0cbb-4138-b9e8-8a6a1bd07128-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "788fb951-0cbb-4138-b9e8-8a6a1bd07128" (UID: "788fb951-0cbb-4138-b9e8-8a6a1bd07128"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 18:45:03 crc kubenswrapper[4808]: I0121 18:45:03.039700 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/788fb951-0cbb-4138-b9e8-8a6a1bd07128-kube-api-access-pl7m6" (OuterVolumeSpecName: "kube-api-access-pl7m6") pod "788fb951-0cbb-4138-b9e8-8a6a1bd07128" (UID: "788fb951-0cbb-4138-b9e8-8a6a1bd07128"). InnerVolumeSpecName "kube-api-access-pl7m6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:45:03 crc kubenswrapper[4808]: I0121 18:45:03.133929 4808 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/788fb951-0cbb-4138-b9e8-8a6a1bd07128-config-volume\") on node \"crc\" DevicePath \"\"" Jan 21 18:45:03 crc kubenswrapper[4808]: I0121 18:45:03.133983 4808 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/788fb951-0cbb-4138-b9e8-8a6a1bd07128-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 21 18:45:03 crc kubenswrapper[4808]: I0121 18:45:03.133997 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pl7m6\" (UniqueName: \"kubernetes.io/projected/788fb951-0cbb-4138-b9e8-8a6a1bd07128-kube-api-access-pl7m6\") on node \"crc\" DevicePath \"\"" Jan 21 18:45:03 crc kubenswrapper[4808]: I0121 18:45:03.741872 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483685-l9l7k" event={"ID":"788fb951-0cbb-4138-b9e8-8a6a1bd07128","Type":"ContainerDied","Data":"719e2442bfc2ea80a88742e652e67d9d6c8ec73bf2a13ced22f6a45f21426512"} Jan 21 18:45:03 crc kubenswrapper[4808]: I0121 18:45:03.741936 4808 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="719e2442bfc2ea80a88742e652e67d9d6c8ec73bf2a13ced22f6a45f21426512" Jan 21 18:45:03 crc kubenswrapper[4808]: I0121 18:45:03.741959 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483685-l9l7k" Jan 21 18:45:04 crc kubenswrapper[4808]: I0121 18:45:04.051382 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q"] Jan 21 18:45:04 crc kubenswrapper[4808]: I0121 18:45:04.060142 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483640-bx86q"] Jan 21 18:45:05 crc kubenswrapper[4808]: I0121 18:45:05.538472 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae9be85d-1887-4d2b-95ad-a71cdd7f8f10" path="/var/lib/kubelet/pods/ae9be85d-1887-4d2b-95ad-a71cdd7f8f10/volumes" Jan 21 18:45:07 crc kubenswrapper[4808]: I0121 18:45:07.468194 4808 scope.go:117] "RemoveContainer" containerID="00c1f4a75e8f3ab0323a0a4797b293d4652edc5c5a12acdfa7fbb9facf6d0c4e" Jan 21 18:45:29 crc kubenswrapper[4808]: I0121 18:45:29.600415 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:45:29 crc kubenswrapper[4808]: I0121 18:45:29.601089 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:45:59 crc kubenswrapper[4808]: I0121 18:45:59.600552 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:45:59 crc kubenswrapper[4808]: I0121 18:45:59.601560 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:45:59 crc kubenswrapper[4808]: I0121 18:45:59.601710 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 18:45:59 crc kubenswrapper[4808]: I0121 18:45:59.602744 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72"} pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 18:45:59 crc kubenswrapper[4808]: I0121 18:45:59.602811 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" containerID="cri-o://dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" gracePeriod=600 Jan 21 18:45:59 crc kubenswrapper[4808]: E0121 18:45:59.743725 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:46:00 crc kubenswrapper[4808]: I0121 18:46:00.138442 4808 generic.go:334] "Generic (PLEG): container finished" podID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" exitCode=0 Jan 21 18:46:00 crc kubenswrapper[4808]: I0121 18:46:00.138497 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerDied","Data":"dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72"} Jan 21 18:46:00 crc kubenswrapper[4808]: I0121 18:46:00.138535 4808 scope.go:117] "RemoveContainer" containerID="a18080ca9142975beeb9b6bd295d2bb48d20c10aa2b2c3996a753b10e7cad3fe" Jan 21 18:46:00 crc kubenswrapper[4808]: I0121 18:46:00.139167 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:46:00 crc kubenswrapper[4808]: E0121 18:46:00.139482 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:46:15 crc kubenswrapper[4808]: I0121 18:46:15.529289 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:46:15 crc kubenswrapper[4808]: E0121 18:46:15.530419 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:46:30 crc kubenswrapper[4808]: I0121 18:46:30.520844 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:46:30 crc kubenswrapper[4808]: E0121 18:46:30.521603 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:46:45 crc kubenswrapper[4808]: I0121 18:46:45.524002 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:46:45 crc kubenswrapper[4808]: E0121 18:46:45.524695 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:46:56 crc kubenswrapper[4808]: I0121 18:46:56.520492 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:46:56 crc kubenswrapper[4808]: E0121 18:46:56.521234 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:47:11 crc kubenswrapper[4808]: I0121 18:47:11.522056 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:47:11 crc kubenswrapper[4808]: E0121 18:47:11.523539 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:47:23 crc kubenswrapper[4808]: I0121 18:47:23.520945 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:47:23 crc kubenswrapper[4808]: E0121 18:47:23.521800 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:47:37 crc kubenswrapper[4808]: I0121 18:47:37.520629 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:47:37 crc kubenswrapper[4808]: E0121 18:47:37.522644 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:47:49 crc kubenswrapper[4808]: I0121 18:47:49.520715 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:47:49 crc kubenswrapper[4808]: E0121 18:47:49.521726 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:48:03 crc kubenswrapper[4808]: I0121 18:48:03.520538 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:48:03 crc kubenswrapper[4808]: E0121 18:48:03.521492 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:48:15 crc kubenswrapper[4808]: I0121 18:48:15.523900 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:48:15 crc kubenswrapper[4808]: E0121 18:48:15.524634 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:48:28 crc kubenswrapper[4808]: I0121 18:48:28.521271 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:48:28 crc kubenswrapper[4808]: E0121 18:48:28.522045 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:48:41 crc kubenswrapper[4808]: I0121 18:48:41.521893 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:48:41 crc kubenswrapper[4808]: E0121 18:48:41.522806 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:48:53 crc kubenswrapper[4808]: I0121 18:48:53.520210 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:48:53 crc kubenswrapper[4808]: E0121 18:48:53.524504 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:49:07 crc kubenswrapper[4808]: I0121 18:49:07.522008 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:49:07 crc kubenswrapper[4808]: E0121 18:49:07.522766 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:49:18 crc kubenswrapper[4808]: I0121 18:49:18.521811 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:49:18 crc kubenswrapper[4808]: E0121 18:49:18.522671 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:49:30 crc kubenswrapper[4808]: I0121 18:49:30.520513 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:49:30 crc kubenswrapper[4808]: E0121 18:49:30.521400 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:49:41 crc kubenswrapper[4808]: I0121 18:49:41.524064 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:49:41 crc kubenswrapper[4808]: E0121 18:49:41.524989 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:49:54 crc kubenswrapper[4808]: I0121 18:49:54.520561 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:49:54 crc kubenswrapper[4808]: E0121 18:49:54.521582 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:50:05 crc kubenswrapper[4808]: I0121 18:50:05.526344 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:50:05 crc kubenswrapper[4808]: E0121 18:50:05.527345 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:50:17 crc kubenswrapper[4808]: I0121 18:50:17.520508 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:50:17 crc kubenswrapper[4808]: E0121 18:50:17.521503 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:50:28 crc kubenswrapper[4808]: I0121 18:50:28.520597 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:50:28 crc kubenswrapper[4808]: E0121 18:50:28.521434 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:50:43 crc kubenswrapper[4808]: I0121 18:50:43.520626 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:50:43 crc kubenswrapper[4808]: E0121 18:50:43.521709 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:50:56 crc kubenswrapper[4808]: I0121 18:50:56.520742 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:50:56 crc kubenswrapper[4808]: E0121 18:50:56.521550 4808 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lgtv9_openshift-machine-config-operator(d76d1c92-28d2-4476-8af9-d95cc779425e)\"" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" Jan 21 18:51:10 crc kubenswrapper[4808]: I0121 18:51:10.521594 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:51:11 crc kubenswrapper[4808]: I0121 18:51:11.222127 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerStarted","Data":"bcebbd6868b78925ce0791cd7c732f3295b97f0561e4f99ba44854c539a50359"} Jan 21 18:53:17 crc kubenswrapper[4808]: I0121 18:53:17.536857 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9x58b"] Jan 21 18:53:17 crc kubenswrapper[4808]: E0121 18:53:17.537595 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788fb951-0cbb-4138-b9e8-8a6a1bd07128" containerName="collect-profiles" Jan 21 18:53:17 crc kubenswrapper[4808]: I0121 18:53:17.537607 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="788fb951-0cbb-4138-b9e8-8a6a1bd07128" containerName="collect-profiles" Jan 21 18:53:17 crc kubenswrapper[4808]: I0121 18:53:17.537717 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="788fb951-0cbb-4138-b9e8-8a6a1bd07128" containerName="collect-profiles" Jan 21 18:53:17 crc kubenswrapper[4808]: I0121 18:53:17.538537 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9x58b" Jan 21 18:53:17 crc kubenswrapper[4808]: I0121 18:53:17.550703 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9x58b"] Jan 21 18:53:17 crc kubenswrapper[4808]: I0121 18:53:17.617482 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e74e774b-8d7d-4b3f-a7fc-c2b316c80900-utilities\") pod \"certified-operators-9x58b\" (UID: \"e74e774b-8d7d-4b3f-a7fc-c2b316c80900\") " pod="openshift-marketplace/certified-operators-9x58b" Jan 21 18:53:17 crc kubenswrapper[4808]: I0121 18:53:17.617677 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e74e774b-8d7d-4b3f-a7fc-c2b316c80900-catalog-content\") pod \"certified-operators-9x58b\" (UID: \"e74e774b-8d7d-4b3f-a7fc-c2b316c80900\") " pod="openshift-marketplace/certified-operators-9x58b" Jan 21 18:53:17 crc kubenswrapper[4808]: I0121 18:53:17.617886 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvvbg\" (UniqueName: \"kubernetes.io/projected/e74e774b-8d7d-4b3f-a7fc-c2b316c80900-kube-api-access-nvvbg\") pod \"certified-operators-9x58b\" (UID: \"e74e774b-8d7d-4b3f-a7fc-c2b316c80900\") " pod="openshift-marketplace/certified-operators-9x58b" Jan 21 18:53:17 crc kubenswrapper[4808]: I0121 18:53:17.718825 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e74e774b-8d7d-4b3f-a7fc-c2b316c80900-utilities\") pod \"certified-operators-9x58b\" (UID: \"e74e774b-8d7d-4b3f-a7fc-c2b316c80900\") " pod="openshift-marketplace/certified-operators-9x58b" Jan 21 18:53:17 crc kubenswrapper[4808]: I0121 18:53:17.718919 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e74e774b-8d7d-4b3f-a7fc-c2b316c80900-catalog-content\") pod \"certified-operators-9x58b\" (UID: \"e74e774b-8d7d-4b3f-a7fc-c2b316c80900\") " pod="openshift-marketplace/certified-operators-9x58b" Jan 21 18:53:17 crc kubenswrapper[4808]: I0121 18:53:17.718968 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvvbg\" (UniqueName: \"kubernetes.io/projected/e74e774b-8d7d-4b3f-a7fc-c2b316c80900-kube-api-access-nvvbg\") pod \"certified-operators-9x58b\" (UID: \"e74e774b-8d7d-4b3f-a7fc-c2b316c80900\") " pod="openshift-marketplace/certified-operators-9x58b" Jan 21 18:53:17 crc kubenswrapper[4808]: I0121 18:53:17.719419 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e74e774b-8d7d-4b3f-a7fc-c2b316c80900-utilities\") pod \"certified-operators-9x58b\" (UID: \"e74e774b-8d7d-4b3f-a7fc-c2b316c80900\") " pod="openshift-marketplace/certified-operators-9x58b" Jan 21 18:53:17 crc kubenswrapper[4808]: I0121 18:53:17.719452 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e74e774b-8d7d-4b3f-a7fc-c2b316c80900-catalog-content\") pod \"certified-operators-9x58b\" (UID: \"e74e774b-8d7d-4b3f-a7fc-c2b316c80900\") " pod="openshift-marketplace/certified-operators-9x58b" Jan 21 18:53:17 crc kubenswrapper[4808]: I0121 18:53:17.749543 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvvbg\" (UniqueName: \"kubernetes.io/projected/e74e774b-8d7d-4b3f-a7fc-c2b316c80900-kube-api-access-nvvbg\") pod \"certified-operators-9x58b\" (UID: \"e74e774b-8d7d-4b3f-a7fc-c2b316c80900\") " pod="openshift-marketplace/certified-operators-9x58b" Jan 21 18:53:17 crc kubenswrapper[4808]: I0121 18:53:17.855062 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9x58b" Jan 21 18:53:18 crc kubenswrapper[4808]: I0121 18:53:18.112283 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9x58b"] Jan 21 18:53:19 crc kubenswrapper[4808]: I0121 18:53:19.080723 4808 generic.go:334] "Generic (PLEG): container finished" podID="e74e774b-8d7d-4b3f-a7fc-c2b316c80900" containerID="944c5d0bb2ef7fd7b2acd15f8bb78cf86e25a5b37f308b98b71e3db384ce23f4" exitCode=0 Jan 21 18:53:19 crc kubenswrapper[4808]: I0121 18:53:19.080824 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x58b" event={"ID":"e74e774b-8d7d-4b3f-a7fc-c2b316c80900","Type":"ContainerDied","Data":"944c5d0bb2ef7fd7b2acd15f8bb78cf86e25a5b37f308b98b71e3db384ce23f4"} Jan 21 18:53:19 crc kubenswrapper[4808]: I0121 18:53:19.081178 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x58b" event={"ID":"e74e774b-8d7d-4b3f-a7fc-c2b316c80900","Type":"ContainerStarted","Data":"c2003e153c154fe4b557efe32a02fd7cd2d58645b0aa92f402b5801b54868a21"} Jan 21 18:53:19 crc kubenswrapper[4808]: I0121 18:53:19.083160 4808 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 18:53:20 crc kubenswrapper[4808]: I0121 18:53:20.088451 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x58b" event={"ID":"e74e774b-8d7d-4b3f-a7fc-c2b316c80900","Type":"ContainerStarted","Data":"0c91a7a59ce6904bbcf710d11f31adf403db5b0deba6465d0ce950477b754077"} Jan 21 18:53:21 crc kubenswrapper[4808]: I0121 18:53:21.096214 4808 generic.go:334] "Generic (PLEG): container finished" podID="e74e774b-8d7d-4b3f-a7fc-c2b316c80900" containerID="0c91a7a59ce6904bbcf710d11f31adf403db5b0deba6465d0ce950477b754077" exitCode=0 Jan 21 18:53:21 crc kubenswrapper[4808]: I0121 18:53:21.096287 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x58b" event={"ID":"e74e774b-8d7d-4b3f-a7fc-c2b316c80900","Type":"ContainerDied","Data":"0c91a7a59ce6904bbcf710d11f31adf403db5b0deba6465d0ce950477b754077"} Jan 21 18:53:22 crc kubenswrapper[4808]: I0121 18:53:22.105217 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x58b" event={"ID":"e74e774b-8d7d-4b3f-a7fc-c2b316c80900","Type":"ContainerStarted","Data":"7857aa4062a0c19906d8141a11cc9b464b4bb7436534f1d7121e09a5a0dd7ca8"} Jan 21 18:53:22 crc kubenswrapper[4808]: I0121 18:53:22.126696 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9x58b" podStartSLOduration=2.724966546 podStartE2EDuration="5.126674233s" podCreationTimestamp="2026-01-21 18:53:17 +0000 UTC" firstStartedPulling="2026-01-21 18:53:19.082893084 +0000 UTC m=+3433.813403969" lastFinishedPulling="2026-01-21 18:53:21.484600781 +0000 UTC m=+3436.215111656" observedRunningTime="2026-01-21 18:53:22.125805991 +0000 UTC m=+3436.856316896" watchObservedRunningTime="2026-01-21 18:53:22.126674233 +0000 UTC m=+3436.857185128" Jan 21 18:53:27 crc kubenswrapper[4808]: I0121 18:53:27.855759 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9x58b" Jan 21 18:53:27 crc kubenswrapper[4808]: I0121 18:53:27.856468 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9x58b" Jan 21 18:53:27 crc kubenswrapper[4808]: I0121 18:53:27.918680 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9x58b" Jan 21 18:53:28 crc kubenswrapper[4808]: I0121 18:53:28.180446 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9x58b" Jan 21 18:53:28 crc kubenswrapper[4808]: I0121 18:53:28.218070 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9x58b"] Jan 21 18:53:29 crc kubenswrapper[4808]: I0121 18:53:29.599959 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:53:29 crc kubenswrapper[4808]: I0121 18:53:29.600026 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:53:30 crc kubenswrapper[4808]: I0121 18:53:30.157918 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9x58b" podUID="e74e774b-8d7d-4b3f-a7fc-c2b316c80900" containerName="registry-server" containerID="cri-o://7857aa4062a0c19906d8141a11cc9b464b4bb7436534f1d7121e09a5a0dd7ca8" gracePeriod=2 Jan 21 18:53:31 crc kubenswrapper[4808]: I0121 18:53:31.612733 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9x58b" Jan 21 18:53:31 crc kubenswrapper[4808]: I0121 18:53:31.709028 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e74e774b-8d7d-4b3f-a7fc-c2b316c80900-catalog-content\") pod \"e74e774b-8d7d-4b3f-a7fc-c2b316c80900\" (UID: \"e74e774b-8d7d-4b3f-a7fc-c2b316c80900\") " Jan 21 18:53:31 crc kubenswrapper[4808]: I0121 18:53:31.722268 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e74e774b-8d7d-4b3f-a7fc-c2b316c80900-utilities\") pod \"e74e774b-8d7d-4b3f-a7fc-c2b316c80900\" (UID: \"e74e774b-8d7d-4b3f-a7fc-c2b316c80900\") " Jan 21 18:53:31 crc kubenswrapper[4808]: I0121 18:53:31.722411 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvvbg\" (UniqueName: \"kubernetes.io/projected/e74e774b-8d7d-4b3f-a7fc-c2b316c80900-kube-api-access-nvvbg\") pod \"e74e774b-8d7d-4b3f-a7fc-c2b316c80900\" (UID: \"e74e774b-8d7d-4b3f-a7fc-c2b316c80900\") " Jan 21 18:53:31 crc kubenswrapper[4808]: I0121 18:53:31.723115 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e74e774b-8d7d-4b3f-a7fc-c2b316c80900-utilities" (OuterVolumeSpecName: "utilities") pod "e74e774b-8d7d-4b3f-a7fc-c2b316c80900" (UID: "e74e774b-8d7d-4b3f-a7fc-c2b316c80900"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:53:31 crc kubenswrapper[4808]: I0121 18:53:31.724875 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e74e774b-8d7d-4b3f-a7fc-c2b316c80900-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:53:31 crc kubenswrapper[4808]: I0121 18:53:31.727483 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e74e774b-8d7d-4b3f-a7fc-c2b316c80900-kube-api-access-nvvbg" (OuterVolumeSpecName: "kube-api-access-nvvbg") pod "e74e774b-8d7d-4b3f-a7fc-c2b316c80900" (UID: "e74e774b-8d7d-4b3f-a7fc-c2b316c80900"). InnerVolumeSpecName "kube-api-access-nvvbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:53:31 crc kubenswrapper[4808]: I0121 18:53:31.767594 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e74e774b-8d7d-4b3f-a7fc-c2b316c80900-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e74e774b-8d7d-4b3f-a7fc-c2b316c80900" (UID: "e74e774b-8d7d-4b3f-a7fc-c2b316c80900"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:53:31 crc kubenswrapper[4808]: I0121 18:53:31.825889 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvvbg\" (UniqueName: \"kubernetes.io/projected/e74e774b-8d7d-4b3f-a7fc-c2b316c80900-kube-api-access-nvvbg\") on node \"crc\" DevicePath \"\"" Jan 21 18:53:31 crc kubenswrapper[4808]: I0121 18:53:31.825921 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e74e774b-8d7d-4b3f-a7fc-c2b316c80900-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:53:32 crc kubenswrapper[4808]: I0121 18:53:32.172453 4808 generic.go:334] "Generic (PLEG): container finished" podID="e74e774b-8d7d-4b3f-a7fc-c2b316c80900" containerID="7857aa4062a0c19906d8141a11cc9b464b4bb7436534f1d7121e09a5a0dd7ca8" exitCode=0 Jan 21 18:53:32 crc kubenswrapper[4808]: I0121 18:53:32.172506 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x58b" event={"ID":"e74e774b-8d7d-4b3f-a7fc-c2b316c80900","Type":"ContainerDied","Data":"7857aa4062a0c19906d8141a11cc9b464b4bb7436534f1d7121e09a5a0dd7ca8"} Jan 21 18:53:32 crc kubenswrapper[4808]: I0121 18:53:32.172526 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9x58b" Jan 21 18:53:32 crc kubenswrapper[4808]: I0121 18:53:32.172543 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x58b" event={"ID":"e74e774b-8d7d-4b3f-a7fc-c2b316c80900","Type":"ContainerDied","Data":"c2003e153c154fe4b557efe32a02fd7cd2d58645b0aa92f402b5801b54868a21"} Jan 21 18:53:32 crc kubenswrapper[4808]: I0121 18:53:32.172566 4808 scope.go:117] "RemoveContainer" containerID="7857aa4062a0c19906d8141a11cc9b464b4bb7436534f1d7121e09a5a0dd7ca8" Jan 21 18:53:32 crc kubenswrapper[4808]: I0121 18:53:32.192234 4808 scope.go:117] "RemoveContainer" containerID="0c91a7a59ce6904bbcf710d11f31adf403db5b0deba6465d0ce950477b754077" Jan 21 18:53:32 crc kubenswrapper[4808]: I0121 18:53:32.211672 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9x58b"] Jan 21 18:53:32 crc kubenswrapper[4808]: I0121 18:53:32.221423 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9x58b"] Jan 21 18:53:32 crc kubenswrapper[4808]: I0121 18:53:32.229012 4808 scope.go:117] "RemoveContainer" containerID="944c5d0bb2ef7fd7b2acd15f8bb78cf86e25a5b37f308b98b71e3db384ce23f4" Jan 21 18:53:32 crc kubenswrapper[4808]: I0121 18:53:32.244652 4808 scope.go:117] "RemoveContainer" containerID="7857aa4062a0c19906d8141a11cc9b464b4bb7436534f1d7121e09a5a0dd7ca8" Jan 21 18:53:32 crc kubenswrapper[4808]: E0121 18:53:32.246624 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7857aa4062a0c19906d8141a11cc9b464b4bb7436534f1d7121e09a5a0dd7ca8\": container with ID starting with 7857aa4062a0c19906d8141a11cc9b464b4bb7436534f1d7121e09a5a0dd7ca8 not found: ID does not exist" containerID="7857aa4062a0c19906d8141a11cc9b464b4bb7436534f1d7121e09a5a0dd7ca8" Jan 21 18:53:32 crc kubenswrapper[4808]: I0121 18:53:32.246673 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7857aa4062a0c19906d8141a11cc9b464b4bb7436534f1d7121e09a5a0dd7ca8"} err="failed to get container status \"7857aa4062a0c19906d8141a11cc9b464b4bb7436534f1d7121e09a5a0dd7ca8\": rpc error: code = NotFound desc = could not find container \"7857aa4062a0c19906d8141a11cc9b464b4bb7436534f1d7121e09a5a0dd7ca8\": container with ID starting with 7857aa4062a0c19906d8141a11cc9b464b4bb7436534f1d7121e09a5a0dd7ca8 not found: ID does not exist" Jan 21 18:53:32 crc kubenswrapper[4808]: I0121 18:53:32.246700 4808 scope.go:117] "RemoveContainer" containerID="0c91a7a59ce6904bbcf710d11f31adf403db5b0deba6465d0ce950477b754077" Jan 21 18:53:32 crc kubenswrapper[4808]: E0121 18:53:32.246976 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c91a7a59ce6904bbcf710d11f31adf403db5b0deba6465d0ce950477b754077\": container with ID starting with 0c91a7a59ce6904bbcf710d11f31adf403db5b0deba6465d0ce950477b754077 not found: ID does not exist" containerID="0c91a7a59ce6904bbcf710d11f31adf403db5b0deba6465d0ce950477b754077" Jan 21 18:53:32 crc kubenswrapper[4808]: I0121 18:53:32.247002 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c91a7a59ce6904bbcf710d11f31adf403db5b0deba6465d0ce950477b754077"} err="failed to get container status \"0c91a7a59ce6904bbcf710d11f31adf403db5b0deba6465d0ce950477b754077\": rpc error: code = NotFound desc = could not find container \"0c91a7a59ce6904bbcf710d11f31adf403db5b0deba6465d0ce950477b754077\": container with ID starting with 0c91a7a59ce6904bbcf710d11f31adf403db5b0deba6465d0ce950477b754077 not found: ID does not exist" Jan 21 18:53:32 crc kubenswrapper[4808]: I0121 18:53:32.247017 4808 scope.go:117] "RemoveContainer" containerID="944c5d0bb2ef7fd7b2acd15f8bb78cf86e25a5b37f308b98b71e3db384ce23f4" Jan 21 18:53:32 crc kubenswrapper[4808]: E0121 18:53:32.247303 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"944c5d0bb2ef7fd7b2acd15f8bb78cf86e25a5b37f308b98b71e3db384ce23f4\": container with ID starting with 944c5d0bb2ef7fd7b2acd15f8bb78cf86e25a5b37f308b98b71e3db384ce23f4 not found: ID does not exist" containerID="944c5d0bb2ef7fd7b2acd15f8bb78cf86e25a5b37f308b98b71e3db384ce23f4" Jan 21 18:53:32 crc kubenswrapper[4808]: I0121 18:53:32.247328 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"944c5d0bb2ef7fd7b2acd15f8bb78cf86e25a5b37f308b98b71e3db384ce23f4"} err="failed to get container status \"944c5d0bb2ef7fd7b2acd15f8bb78cf86e25a5b37f308b98b71e3db384ce23f4\": rpc error: code = NotFound desc = could not find container \"944c5d0bb2ef7fd7b2acd15f8bb78cf86e25a5b37f308b98b71e3db384ce23f4\": container with ID starting with 944c5d0bb2ef7fd7b2acd15f8bb78cf86e25a5b37f308b98b71e3db384ce23f4 not found: ID does not exist" Jan 21 18:53:33 crc kubenswrapper[4808]: I0121 18:53:33.537740 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e74e774b-8d7d-4b3f-a7fc-c2b316c80900" path="/var/lib/kubelet/pods/e74e774b-8d7d-4b3f-a7fc-c2b316c80900/volumes" Jan 21 18:53:59 crc kubenswrapper[4808]: I0121 18:53:59.600046 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:53:59 crc kubenswrapper[4808]: I0121 18:53:59.601945 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:54:17 crc kubenswrapper[4808]: I0121 18:54:17.786072 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t76pw"] Jan 21 18:54:17 crc kubenswrapper[4808]: E0121 18:54:17.787833 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e74e774b-8d7d-4b3f-a7fc-c2b316c80900" containerName="extract-utilities" Jan 21 18:54:17 crc kubenswrapper[4808]: I0121 18:54:17.787855 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e74e774b-8d7d-4b3f-a7fc-c2b316c80900" containerName="extract-utilities" Jan 21 18:54:17 crc kubenswrapper[4808]: E0121 18:54:17.787871 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e74e774b-8d7d-4b3f-a7fc-c2b316c80900" containerName="extract-content" Jan 21 18:54:17 crc kubenswrapper[4808]: I0121 18:54:17.787878 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e74e774b-8d7d-4b3f-a7fc-c2b316c80900" containerName="extract-content" Jan 21 18:54:17 crc kubenswrapper[4808]: E0121 18:54:17.788651 4808 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e74e774b-8d7d-4b3f-a7fc-c2b316c80900" containerName="registry-server" Jan 21 18:54:17 crc kubenswrapper[4808]: I0121 18:54:17.788674 4808 state_mem.go:107] "Deleted CPUSet assignment" podUID="e74e774b-8d7d-4b3f-a7fc-c2b316c80900" containerName="registry-server" Jan 21 18:54:17 crc kubenswrapper[4808]: I0121 18:54:17.788834 4808 memory_manager.go:354] "RemoveStaleState removing state" podUID="e74e774b-8d7d-4b3f-a7fc-c2b316c80900" containerName="registry-server" Jan 21 18:54:17 crc kubenswrapper[4808]: I0121 18:54:17.790491 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t76pw" Jan 21 18:54:17 crc kubenswrapper[4808]: I0121 18:54:17.807927 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t76pw"] Jan 21 18:54:17 crc kubenswrapper[4808]: I0121 18:54:17.878469 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be6c83b7-a19e-48cf-9775-1d0fccb76888-utilities\") pod \"redhat-operators-t76pw\" (UID: \"be6c83b7-a19e-48cf-9775-1d0fccb76888\") " pod="openshift-marketplace/redhat-operators-t76pw" Jan 21 18:54:17 crc kubenswrapper[4808]: I0121 18:54:17.878520 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be6c83b7-a19e-48cf-9775-1d0fccb76888-catalog-content\") pod \"redhat-operators-t76pw\" (UID: \"be6c83b7-a19e-48cf-9775-1d0fccb76888\") " pod="openshift-marketplace/redhat-operators-t76pw" Jan 21 18:54:17 crc kubenswrapper[4808]: I0121 18:54:17.878586 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wsnc\" (UniqueName: \"kubernetes.io/projected/be6c83b7-a19e-48cf-9775-1d0fccb76888-kube-api-access-7wsnc\") pod \"redhat-operators-t76pw\" (UID: \"be6c83b7-a19e-48cf-9775-1d0fccb76888\") " pod="openshift-marketplace/redhat-operators-t76pw" Jan 21 18:54:17 crc kubenswrapper[4808]: I0121 18:54:17.980768 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be6c83b7-a19e-48cf-9775-1d0fccb76888-utilities\") pod \"redhat-operators-t76pw\" (UID: \"be6c83b7-a19e-48cf-9775-1d0fccb76888\") " pod="openshift-marketplace/redhat-operators-t76pw" Jan 21 18:54:17 crc kubenswrapper[4808]: I0121 18:54:17.980822 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be6c83b7-a19e-48cf-9775-1d0fccb76888-catalog-content\") pod \"redhat-operators-t76pw\" (UID: \"be6c83b7-a19e-48cf-9775-1d0fccb76888\") " pod="openshift-marketplace/redhat-operators-t76pw" Jan 21 18:54:17 crc kubenswrapper[4808]: I0121 18:54:17.980871 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wsnc\" (UniqueName: \"kubernetes.io/projected/be6c83b7-a19e-48cf-9775-1d0fccb76888-kube-api-access-7wsnc\") pod \"redhat-operators-t76pw\" (UID: \"be6c83b7-a19e-48cf-9775-1d0fccb76888\") " pod="openshift-marketplace/redhat-operators-t76pw" Jan 21 18:54:17 crc kubenswrapper[4808]: I0121 18:54:17.981380 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be6c83b7-a19e-48cf-9775-1d0fccb76888-utilities\") pod \"redhat-operators-t76pw\" (UID: \"be6c83b7-a19e-48cf-9775-1d0fccb76888\") " pod="openshift-marketplace/redhat-operators-t76pw" Jan 21 18:54:17 crc kubenswrapper[4808]: I0121 18:54:17.981590 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be6c83b7-a19e-48cf-9775-1d0fccb76888-catalog-content\") pod \"redhat-operators-t76pw\" (UID: \"be6c83b7-a19e-48cf-9775-1d0fccb76888\") " pod="openshift-marketplace/redhat-operators-t76pw" Jan 21 18:54:18 crc kubenswrapper[4808]: I0121 18:54:18.007324 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wsnc\" (UniqueName: \"kubernetes.io/projected/be6c83b7-a19e-48cf-9775-1d0fccb76888-kube-api-access-7wsnc\") pod \"redhat-operators-t76pw\" (UID: \"be6c83b7-a19e-48cf-9775-1d0fccb76888\") " pod="openshift-marketplace/redhat-operators-t76pw" Jan 21 18:54:18 crc kubenswrapper[4808]: I0121 18:54:18.122847 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t76pw" Jan 21 18:54:18 crc kubenswrapper[4808]: I0121 18:54:18.328690 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t76pw"] Jan 21 18:54:18 crc kubenswrapper[4808]: I0121 18:54:18.470854 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t76pw" event={"ID":"be6c83b7-a19e-48cf-9775-1d0fccb76888","Type":"ContainerStarted","Data":"b21bf2c65ed651724428d47e116da14283298bf52e48053427a5c86bc06760f2"} Jan 21 18:54:19 crc kubenswrapper[4808]: I0121 18:54:19.478133 4808 generic.go:334] "Generic (PLEG): container finished" podID="be6c83b7-a19e-48cf-9775-1d0fccb76888" containerID="4d695677ac85d882c6f9584ab4ece9da3cc0632ab42714622800ac645c6af75b" exitCode=0 Jan 21 18:54:19 crc kubenswrapper[4808]: I0121 18:54:19.478430 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t76pw" event={"ID":"be6c83b7-a19e-48cf-9775-1d0fccb76888","Type":"ContainerDied","Data":"4d695677ac85d882c6f9584ab4ece9da3cc0632ab42714622800ac645c6af75b"} Jan 21 18:54:21 crc kubenswrapper[4808]: I0121 18:54:21.494581 4808 generic.go:334] "Generic (PLEG): container finished" podID="be6c83b7-a19e-48cf-9775-1d0fccb76888" containerID="5faafab365769168079ff9cb7b48530dd282372405c65fced5ae5b72586983de" exitCode=0 Jan 21 18:54:21 crc kubenswrapper[4808]: I0121 18:54:21.494657 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t76pw" event={"ID":"be6c83b7-a19e-48cf-9775-1d0fccb76888","Type":"ContainerDied","Data":"5faafab365769168079ff9cb7b48530dd282372405c65fced5ae5b72586983de"} Jan 21 18:54:22 crc kubenswrapper[4808]: I0121 18:54:22.505608 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t76pw" event={"ID":"be6c83b7-a19e-48cf-9775-1d0fccb76888","Type":"ContainerStarted","Data":"09758bdf079f74688fdfa1a8a2fd4b9f105cc287264b8b1587a7ad09b12be3e3"} Jan 21 18:54:22 crc kubenswrapper[4808]: I0121 18:54:22.528005 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t76pw" podStartSLOduration=3.098674132 podStartE2EDuration="5.527988066s" podCreationTimestamp="2026-01-21 18:54:17 +0000 UTC" firstStartedPulling="2026-01-21 18:54:19.480479806 +0000 UTC m=+3494.210990691" lastFinishedPulling="2026-01-21 18:54:21.90979374 +0000 UTC m=+3496.640304625" observedRunningTime="2026-01-21 18:54:22.523336261 +0000 UTC m=+3497.253847166" watchObservedRunningTime="2026-01-21 18:54:22.527988066 +0000 UTC m=+3497.258498941" Jan 21 18:54:25 crc kubenswrapper[4808]: I0121 18:54:25.575813 4808 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wd78g"] Jan 21 18:54:25 crc kubenswrapper[4808]: I0121 18:54:25.577771 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wd78g" Jan 21 18:54:25 crc kubenswrapper[4808]: I0121 18:54:25.595310 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wd78g"] Jan 21 18:54:25 crc kubenswrapper[4808]: I0121 18:54:25.687312 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc4d25a5-12f9-418c-902e-4ea792a5c7c1-catalog-content\") pod \"community-operators-wd78g\" (UID: \"cc4d25a5-12f9-418c-902e-4ea792a5c7c1\") " pod="openshift-marketplace/community-operators-wd78g" Jan 21 18:54:25 crc kubenswrapper[4808]: I0121 18:54:25.687395 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj7rl\" (UniqueName: \"kubernetes.io/projected/cc4d25a5-12f9-418c-902e-4ea792a5c7c1-kube-api-access-mj7rl\") pod \"community-operators-wd78g\" (UID: \"cc4d25a5-12f9-418c-902e-4ea792a5c7c1\") " pod="openshift-marketplace/community-operators-wd78g" Jan 21 18:54:25 crc kubenswrapper[4808]: I0121 18:54:25.687423 4808 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc4d25a5-12f9-418c-902e-4ea792a5c7c1-utilities\") pod \"community-operators-wd78g\" (UID: \"cc4d25a5-12f9-418c-902e-4ea792a5c7c1\") " pod="openshift-marketplace/community-operators-wd78g" Jan 21 18:54:25 crc kubenswrapper[4808]: I0121 18:54:25.789329 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj7rl\" (UniqueName: \"kubernetes.io/projected/cc4d25a5-12f9-418c-902e-4ea792a5c7c1-kube-api-access-mj7rl\") pod \"community-operators-wd78g\" (UID: \"cc4d25a5-12f9-418c-902e-4ea792a5c7c1\") " pod="openshift-marketplace/community-operators-wd78g" Jan 21 18:54:25 crc kubenswrapper[4808]: I0121 18:54:25.789436 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc4d25a5-12f9-418c-902e-4ea792a5c7c1-utilities\") pod \"community-operators-wd78g\" (UID: \"cc4d25a5-12f9-418c-902e-4ea792a5c7c1\") " pod="openshift-marketplace/community-operators-wd78g" Jan 21 18:54:25 crc kubenswrapper[4808]: I0121 18:54:25.789512 4808 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc4d25a5-12f9-418c-902e-4ea792a5c7c1-catalog-content\") pod \"community-operators-wd78g\" (UID: \"cc4d25a5-12f9-418c-902e-4ea792a5c7c1\") " pod="openshift-marketplace/community-operators-wd78g" Jan 21 18:54:25 crc kubenswrapper[4808]: I0121 18:54:25.790032 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc4d25a5-12f9-418c-902e-4ea792a5c7c1-utilities\") pod \"community-operators-wd78g\" (UID: \"cc4d25a5-12f9-418c-902e-4ea792a5c7c1\") " pod="openshift-marketplace/community-operators-wd78g" Jan 21 18:54:25 crc kubenswrapper[4808]: I0121 18:54:25.790137 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc4d25a5-12f9-418c-902e-4ea792a5c7c1-catalog-content\") pod \"community-operators-wd78g\" (UID: \"cc4d25a5-12f9-418c-902e-4ea792a5c7c1\") " pod="openshift-marketplace/community-operators-wd78g" Jan 21 18:54:25 crc kubenswrapper[4808]: I0121 18:54:25.814411 4808 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj7rl\" (UniqueName: \"kubernetes.io/projected/cc4d25a5-12f9-418c-902e-4ea792a5c7c1-kube-api-access-mj7rl\") pod \"community-operators-wd78g\" (UID: \"cc4d25a5-12f9-418c-902e-4ea792a5c7c1\") " pod="openshift-marketplace/community-operators-wd78g" Jan 21 18:54:25 crc kubenswrapper[4808]: I0121 18:54:25.900063 4808 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wd78g" Jan 21 18:54:26 crc kubenswrapper[4808]: I0121 18:54:26.260301 4808 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wd78g"] Jan 21 18:54:26 crc kubenswrapper[4808]: I0121 18:54:26.532450 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd78g" event={"ID":"cc4d25a5-12f9-418c-902e-4ea792a5c7c1","Type":"ContainerStarted","Data":"6b7488f9166dbddd8befb5ae5d60fd3eab9b76969564e60d717477e8e5ab6c36"} Jan 21 18:54:28 crc kubenswrapper[4808]: I0121 18:54:28.123518 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t76pw" Jan 21 18:54:28 crc kubenswrapper[4808]: I0121 18:54:28.124954 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t76pw" Jan 21 18:54:28 crc kubenswrapper[4808]: I0121 18:54:28.168278 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t76pw" Jan 21 18:54:28 crc kubenswrapper[4808]: I0121 18:54:28.544996 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd78g" event={"ID":"cc4d25a5-12f9-418c-902e-4ea792a5c7c1","Type":"ContainerStarted","Data":"6269ae7e917f9f279ff59fe5bff85ccc50c2936185e6e48a2adb6342f440d823"} Jan 21 18:54:28 crc kubenswrapper[4808]: I0121 18:54:28.589674 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t76pw" Jan 21 18:54:29 crc kubenswrapper[4808]: I0121 18:54:29.552610 4808 generic.go:334] "Generic (PLEG): container finished" podID="cc4d25a5-12f9-418c-902e-4ea792a5c7c1" containerID="6269ae7e917f9f279ff59fe5bff85ccc50c2936185e6e48a2adb6342f440d823" exitCode=0 Jan 21 18:54:29 crc kubenswrapper[4808]: I0121 18:54:29.553531 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd78g" event={"ID":"cc4d25a5-12f9-418c-902e-4ea792a5c7c1","Type":"ContainerDied","Data":"6269ae7e917f9f279ff59fe5bff85ccc50c2936185e6e48a2adb6342f440d823"} Jan 21 18:54:29 crc kubenswrapper[4808]: I0121 18:54:29.571483 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t76pw"] Jan 21 18:54:29 crc kubenswrapper[4808]: I0121 18:54:29.599928 4808 patch_prober.go:28] interesting pod/machine-config-daemon-lgtv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 18:54:29 crc kubenswrapper[4808]: I0121 18:54:29.600026 4808 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 18:54:29 crc kubenswrapper[4808]: I0121 18:54:29.600097 4808 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" Jan 21 18:54:29 crc kubenswrapper[4808]: I0121 18:54:29.600972 4808 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bcebbd6868b78925ce0791cd7c732f3295b97f0561e4f99ba44854c539a50359"} pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 18:54:29 crc kubenswrapper[4808]: I0121 18:54:29.601045 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" podUID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerName="machine-config-daemon" containerID="cri-o://bcebbd6868b78925ce0791cd7c732f3295b97f0561e4f99ba44854c539a50359" gracePeriod=600 Jan 21 18:54:30 crc kubenswrapper[4808]: I0121 18:54:30.563278 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd78g" event={"ID":"cc4d25a5-12f9-418c-902e-4ea792a5c7c1","Type":"ContainerStarted","Data":"75dcae6c6ce7b4c73af19aca2f9c39f91576dc3e329abcd5ba49cea177a3e9d2"} Jan 21 18:54:30 crc kubenswrapper[4808]: I0121 18:54:30.566045 4808 generic.go:334] "Generic (PLEG): container finished" podID="d76d1c92-28d2-4476-8af9-d95cc779425e" containerID="bcebbd6868b78925ce0791cd7c732f3295b97f0561e4f99ba44854c539a50359" exitCode=0 Jan 21 18:54:30 crc kubenswrapper[4808]: I0121 18:54:30.566150 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerDied","Data":"bcebbd6868b78925ce0791cd7c732f3295b97f0561e4f99ba44854c539a50359"} Jan 21 18:54:30 crc kubenswrapper[4808]: I0121 18:54:30.566802 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lgtv9" event={"ID":"d76d1c92-28d2-4476-8af9-d95cc779425e","Type":"ContainerStarted","Data":"7321ff3599270edff67366ed2b8a5fe17d50a608c0ed8aa7c63d3c97d96d0e4a"} Jan 21 18:54:30 crc kubenswrapper[4808]: I0121 18:54:30.566827 4808 scope.go:117] "RemoveContainer" containerID="dbb181f188385169ef4641f1280611eeed8e318e2b8f6d78131a7d83d7e17b72" Jan 21 18:54:31 crc kubenswrapper[4808]: I0121 18:54:31.577681 4808 generic.go:334] "Generic (PLEG): container finished" podID="cc4d25a5-12f9-418c-902e-4ea792a5c7c1" containerID="75dcae6c6ce7b4c73af19aca2f9c39f91576dc3e329abcd5ba49cea177a3e9d2" exitCode=0 Jan 21 18:54:31 crc kubenswrapper[4808]: I0121 18:54:31.578496 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd78g" event={"ID":"cc4d25a5-12f9-418c-902e-4ea792a5c7c1","Type":"ContainerDied","Data":"75dcae6c6ce7b4c73af19aca2f9c39f91576dc3e329abcd5ba49cea177a3e9d2"} Jan 21 18:54:31 crc kubenswrapper[4808]: I0121 18:54:31.586855 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t76pw" podUID="be6c83b7-a19e-48cf-9775-1d0fccb76888" containerName="registry-server" containerID="cri-o://09758bdf079f74688fdfa1a8a2fd4b9f105cc287264b8b1587a7ad09b12be3e3" gracePeriod=2 Jan 21 18:54:34 crc kubenswrapper[4808]: I0121 18:54:34.606398 4808 generic.go:334] "Generic (PLEG): container finished" podID="be6c83b7-a19e-48cf-9775-1d0fccb76888" containerID="09758bdf079f74688fdfa1a8a2fd4b9f105cc287264b8b1587a7ad09b12be3e3" exitCode=0 Jan 21 18:54:34 crc kubenswrapper[4808]: I0121 18:54:34.606442 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t76pw" event={"ID":"be6c83b7-a19e-48cf-9775-1d0fccb76888","Type":"ContainerDied","Data":"09758bdf079f74688fdfa1a8a2fd4b9f105cc287264b8b1587a7ad09b12be3e3"} Jan 21 18:54:34 crc kubenswrapper[4808]: I0121 18:54:34.608937 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd78g" event={"ID":"cc4d25a5-12f9-418c-902e-4ea792a5c7c1","Type":"ContainerStarted","Data":"549b7319d1f0d84d0c37092caef3a155144a9a8e78349e0b195a5e9cd66835f1"} Jan 21 18:54:34 crc kubenswrapper[4808]: I0121 18:54:34.626525 4808 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wd78g" podStartSLOduration=5.375908024 podStartE2EDuration="9.626501466s" podCreationTimestamp="2026-01-21 18:54:25 +0000 UTC" firstStartedPulling="2026-01-21 18:54:29.5540591 +0000 UTC m=+3504.284569985" lastFinishedPulling="2026-01-21 18:54:33.804652532 +0000 UTC m=+3508.535163427" observedRunningTime="2026-01-21 18:54:34.626128397 +0000 UTC m=+3509.356639282" watchObservedRunningTime="2026-01-21 18:54:34.626501466 +0000 UTC m=+3509.357012361" Jan 21 18:54:34 crc kubenswrapper[4808]: I0121 18:54:34.724346 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t76pw" Jan 21 18:54:34 crc kubenswrapper[4808]: I0121 18:54:34.820963 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wsnc\" (UniqueName: \"kubernetes.io/projected/be6c83b7-a19e-48cf-9775-1d0fccb76888-kube-api-access-7wsnc\") pod \"be6c83b7-a19e-48cf-9775-1d0fccb76888\" (UID: \"be6c83b7-a19e-48cf-9775-1d0fccb76888\") " Jan 21 18:54:34 crc kubenswrapper[4808]: I0121 18:54:34.821102 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be6c83b7-a19e-48cf-9775-1d0fccb76888-utilities\") pod \"be6c83b7-a19e-48cf-9775-1d0fccb76888\" (UID: \"be6c83b7-a19e-48cf-9775-1d0fccb76888\") " Jan 21 18:54:34 crc kubenswrapper[4808]: I0121 18:54:34.821136 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be6c83b7-a19e-48cf-9775-1d0fccb76888-catalog-content\") pod \"be6c83b7-a19e-48cf-9775-1d0fccb76888\" (UID: \"be6c83b7-a19e-48cf-9775-1d0fccb76888\") " Jan 21 18:54:34 crc kubenswrapper[4808]: I0121 18:54:34.822393 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be6c83b7-a19e-48cf-9775-1d0fccb76888-utilities" (OuterVolumeSpecName: "utilities") pod "be6c83b7-a19e-48cf-9775-1d0fccb76888" (UID: "be6c83b7-a19e-48cf-9775-1d0fccb76888"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:54:34 crc kubenswrapper[4808]: I0121 18:54:34.830281 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be6c83b7-a19e-48cf-9775-1d0fccb76888-kube-api-access-7wsnc" (OuterVolumeSpecName: "kube-api-access-7wsnc") pod "be6c83b7-a19e-48cf-9775-1d0fccb76888" (UID: "be6c83b7-a19e-48cf-9775-1d0fccb76888"). InnerVolumeSpecName "kube-api-access-7wsnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:54:34 crc kubenswrapper[4808]: I0121 18:54:34.922872 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wsnc\" (UniqueName: \"kubernetes.io/projected/be6c83b7-a19e-48cf-9775-1d0fccb76888-kube-api-access-7wsnc\") on node \"crc\" DevicePath \"\"" Jan 21 18:54:34 crc kubenswrapper[4808]: I0121 18:54:34.922907 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be6c83b7-a19e-48cf-9775-1d0fccb76888-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:54:34 crc kubenswrapper[4808]: I0121 18:54:34.954614 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be6c83b7-a19e-48cf-9775-1d0fccb76888-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be6c83b7-a19e-48cf-9775-1d0fccb76888" (UID: "be6c83b7-a19e-48cf-9775-1d0fccb76888"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:54:35 crc kubenswrapper[4808]: I0121 18:54:35.024174 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be6c83b7-a19e-48cf-9775-1d0fccb76888-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:54:35 crc kubenswrapper[4808]: I0121 18:54:35.618502 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t76pw" event={"ID":"be6c83b7-a19e-48cf-9775-1d0fccb76888","Type":"ContainerDied","Data":"b21bf2c65ed651724428d47e116da14283298bf52e48053427a5c86bc06760f2"} Jan 21 18:54:35 crc kubenswrapper[4808]: I0121 18:54:35.618564 4808 scope.go:117] "RemoveContainer" containerID="09758bdf079f74688fdfa1a8a2fd4b9f105cc287264b8b1587a7ad09b12be3e3" Jan 21 18:54:35 crc kubenswrapper[4808]: I0121 18:54:35.618869 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t76pw" Jan 21 18:54:35 crc kubenswrapper[4808]: I0121 18:54:35.638305 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t76pw"] Jan 21 18:54:35 crc kubenswrapper[4808]: I0121 18:54:35.640169 4808 scope.go:117] "RemoveContainer" containerID="5faafab365769168079ff9cb7b48530dd282372405c65fced5ae5b72586983de" Jan 21 18:54:35 crc kubenswrapper[4808]: I0121 18:54:35.647595 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t76pw"] Jan 21 18:54:35 crc kubenswrapper[4808]: I0121 18:54:35.661255 4808 scope.go:117] "RemoveContainer" containerID="4d695677ac85d882c6f9584ab4ece9da3cc0632ab42714622800ac645c6af75b" Jan 21 18:54:35 crc kubenswrapper[4808]: I0121 18:54:35.900713 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wd78g" Jan 21 18:54:35 crc kubenswrapper[4808]: I0121 18:54:35.900934 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wd78g" Jan 21 18:54:35 crc kubenswrapper[4808]: I0121 18:54:35.957175 4808 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wd78g" Jan 21 18:54:37 crc kubenswrapper[4808]: I0121 18:54:37.529115 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be6c83b7-a19e-48cf-9775-1d0fccb76888" path="/var/lib/kubelet/pods/be6c83b7-a19e-48cf-9775-1d0fccb76888/volumes" Jan 21 18:54:45 crc kubenswrapper[4808]: I0121 18:54:45.950597 4808 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wd78g" Jan 21 18:54:46 crc kubenswrapper[4808]: I0121 18:54:46.010440 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wd78g"] Jan 21 18:54:46 crc kubenswrapper[4808]: I0121 18:54:46.685195 4808 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wd78g" podUID="cc4d25a5-12f9-418c-902e-4ea792a5c7c1" containerName="registry-server" containerID="cri-o://549b7319d1f0d84d0c37092caef3a155144a9a8e78349e0b195a5e9cd66835f1" gracePeriod=2 Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.044021 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wd78g" Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.211984 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc4d25a5-12f9-418c-902e-4ea792a5c7c1-catalog-content\") pod \"cc4d25a5-12f9-418c-902e-4ea792a5c7c1\" (UID: \"cc4d25a5-12f9-418c-902e-4ea792a5c7c1\") " Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.212052 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mj7rl\" (UniqueName: \"kubernetes.io/projected/cc4d25a5-12f9-418c-902e-4ea792a5c7c1-kube-api-access-mj7rl\") pod \"cc4d25a5-12f9-418c-902e-4ea792a5c7c1\" (UID: \"cc4d25a5-12f9-418c-902e-4ea792a5c7c1\") " Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.212107 4808 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc4d25a5-12f9-418c-902e-4ea792a5c7c1-utilities\") pod \"cc4d25a5-12f9-418c-902e-4ea792a5c7c1\" (UID: \"cc4d25a5-12f9-418c-902e-4ea792a5c7c1\") " Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.213162 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc4d25a5-12f9-418c-902e-4ea792a5c7c1-utilities" (OuterVolumeSpecName: "utilities") pod "cc4d25a5-12f9-418c-902e-4ea792a5c7c1" (UID: "cc4d25a5-12f9-418c-902e-4ea792a5c7c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.219465 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc4d25a5-12f9-418c-902e-4ea792a5c7c1-kube-api-access-mj7rl" (OuterVolumeSpecName: "kube-api-access-mj7rl") pod "cc4d25a5-12f9-418c-902e-4ea792a5c7c1" (UID: "cc4d25a5-12f9-418c-902e-4ea792a5c7c1"). InnerVolumeSpecName "kube-api-access-mj7rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.264656 4808 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc4d25a5-12f9-418c-902e-4ea792a5c7c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc4d25a5-12f9-418c-902e-4ea792a5c7c1" (UID: "cc4d25a5-12f9-418c-902e-4ea792a5c7c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.320862 4808 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc4d25a5-12f9-418c-902e-4ea792a5c7c1-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.321277 4808 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mj7rl\" (UniqueName: \"kubernetes.io/projected/cc4d25a5-12f9-418c-902e-4ea792a5c7c1-kube-api-access-mj7rl\") on node \"crc\" DevicePath \"\"" Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.321376 4808 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc4d25a5-12f9-418c-902e-4ea792a5c7c1-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.693297 4808 generic.go:334] "Generic (PLEG): container finished" podID="cc4d25a5-12f9-418c-902e-4ea792a5c7c1" containerID="549b7319d1f0d84d0c37092caef3a155144a9a8e78349e0b195a5e9cd66835f1" exitCode=0 Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.693554 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd78g" event={"ID":"cc4d25a5-12f9-418c-902e-4ea792a5c7c1","Type":"ContainerDied","Data":"549b7319d1f0d84d0c37092caef3a155144a9a8e78349e0b195a5e9cd66835f1"} Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.693678 4808 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd78g" event={"ID":"cc4d25a5-12f9-418c-902e-4ea792a5c7c1","Type":"ContainerDied","Data":"6b7488f9166dbddd8befb5ae5d60fd3eab9b76969564e60d717477e8e5ab6c36"} Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.693771 4808 scope.go:117] "RemoveContainer" containerID="549b7319d1f0d84d0c37092caef3a155144a9a8e78349e0b195a5e9cd66835f1" Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.693973 4808 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wd78g" Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.717644 4808 scope.go:117] "RemoveContainer" containerID="75dcae6c6ce7b4c73af19aca2f9c39f91576dc3e329abcd5ba49cea177a3e9d2" Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.718575 4808 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wd78g"] Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.723218 4808 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wd78g"] Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.735917 4808 scope.go:117] "RemoveContainer" containerID="6269ae7e917f9f279ff59fe5bff85ccc50c2936185e6e48a2adb6342f440d823" Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.756274 4808 scope.go:117] "RemoveContainer" containerID="549b7319d1f0d84d0c37092caef3a155144a9a8e78349e0b195a5e9cd66835f1" Jan 21 18:54:47 crc kubenswrapper[4808]: E0121 18:54:47.756746 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"549b7319d1f0d84d0c37092caef3a155144a9a8e78349e0b195a5e9cd66835f1\": container with ID starting with 549b7319d1f0d84d0c37092caef3a155144a9a8e78349e0b195a5e9cd66835f1 not found: ID does not exist" containerID="549b7319d1f0d84d0c37092caef3a155144a9a8e78349e0b195a5e9cd66835f1" Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.756835 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"549b7319d1f0d84d0c37092caef3a155144a9a8e78349e0b195a5e9cd66835f1"} err="failed to get container status \"549b7319d1f0d84d0c37092caef3a155144a9a8e78349e0b195a5e9cd66835f1\": rpc error: code = NotFound desc = could not find container \"549b7319d1f0d84d0c37092caef3a155144a9a8e78349e0b195a5e9cd66835f1\": container with ID starting with 549b7319d1f0d84d0c37092caef3a155144a9a8e78349e0b195a5e9cd66835f1 not found: ID does not exist" Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.756865 4808 scope.go:117] "RemoveContainer" containerID="75dcae6c6ce7b4c73af19aca2f9c39f91576dc3e329abcd5ba49cea177a3e9d2" Jan 21 18:54:47 crc kubenswrapper[4808]: E0121 18:54:47.757210 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75dcae6c6ce7b4c73af19aca2f9c39f91576dc3e329abcd5ba49cea177a3e9d2\": container with ID starting with 75dcae6c6ce7b4c73af19aca2f9c39f91576dc3e329abcd5ba49cea177a3e9d2 not found: ID does not exist" containerID="75dcae6c6ce7b4c73af19aca2f9c39f91576dc3e329abcd5ba49cea177a3e9d2" Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.757275 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75dcae6c6ce7b4c73af19aca2f9c39f91576dc3e329abcd5ba49cea177a3e9d2"} err="failed to get container status \"75dcae6c6ce7b4c73af19aca2f9c39f91576dc3e329abcd5ba49cea177a3e9d2\": rpc error: code = NotFound desc = could not find container \"75dcae6c6ce7b4c73af19aca2f9c39f91576dc3e329abcd5ba49cea177a3e9d2\": container with ID starting with 75dcae6c6ce7b4c73af19aca2f9c39f91576dc3e329abcd5ba49cea177a3e9d2 not found: ID does not exist" Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.757298 4808 scope.go:117] "RemoveContainer" containerID="6269ae7e917f9f279ff59fe5bff85ccc50c2936185e6e48a2adb6342f440d823" Jan 21 18:54:47 crc kubenswrapper[4808]: E0121 18:54:47.757515 4808 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6269ae7e917f9f279ff59fe5bff85ccc50c2936185e6e48a2adb6342f440d823\": container with ID starting with 6269ae7e917f9f279ff59fe5bff85ccc50c2936185e6e48a2adb6342f440d823 not found: ID does not exist" containerID="6269ae7e917f9f279ff59fe5bff85ccc50c2936185e6e48a2adb6342f440d823" Jan 21 18:54:47 crc kubenswrapper[4808]: I0121 18:54:47.757553 4808 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6269ae7e917f9f279ff59fe5bff85ccc50c2936185e6e48a2adb6342f440d823"} err="failed to get container status \"6269ae7e917f9f279ff59fe5bff85ccc50c2936185e6e48a2adb6342f440d823\": rpc error: code = NotFound desc = could not find container \"6269ae7e917f9f279ff59fe5bff85ccc50c2936185e6e48a2adb6342f440d823\": container with ID starting with 6269ae7e917f9f279ff59fe5bff85ccc50c2936185e6e48a2adb6342f440d823 not found: ID does not exist" Jan 21 18:54:49 crc kubenswrapper[4808]: I0121 18:54:49.528893 4808 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc4d25a5-12f9-418c-902e-4ea792a5c7c1" path="/var/lib/kubelet/pods/cc4d25a5-12f9-418c-902e-4ea792a5c7c1/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515134220415024442 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015134220416017360 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015134211121016474 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015134211121015444 5ustar corecore